code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) A = { """configuration_owlvit""": [ """OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """OwlViTConfig""", """OwlViTOnnxConfig""", """OwlViTTextConfig""", """OwlViTVisionConfig""", ], """processing_owlvit""": ["""OwlViTProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""OwlViTFeatureExtractor"""] A = ["""OwlViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """OwlViTModel""", """OwlViTPreTrainedModel""", """OwlViTTextModel""", """OwlViTVisionModel""", """OwlViTForObjectDetection""", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __UpperCAmelCase : Optional[Any] = TapasConfig.from_json_file(UpperCamelCase ) # set absolute/relative position embeddings parameter __UpperCAmelCase : Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __UpperCAmelCase : List[str] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WTQ": # run_task_main.py hparams __UpperCAmelCase : Tuple = 4 __UpperCAmelCase : Any = True # hparam_utils.py hparams __UpperCAmelCase : Union[str, Any] = 0.664694 __UpperCAmelCase : Union[str, Any] = 0.207951 __UpperCAmelCase : int = 0.121194 __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : List[str] = True __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : List[str] = 0.0352513 __UpperCAmelCase : Optional[int] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __UpperCAmelCase : int = 4 __UpperCAmelCase : Optional[int] = False # hparam_utils.py hparams __UpperCAmelCase : int = 36.4519 __UpperCAmelCase : str = 0.903421 __UpperCAmelCase : Dict = 222.088 __UpperCAmelCase : Dict = True __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Tuple = True __UpperCAmelCase : Any = 0.763141 __UpperCAmelCase : Optional[Any] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "TABFACT": __UpperCAmelCase : Union[str, Any] = TapasForSequenceClassification(config=UpperCamelCase ) elif task == "MLM": __UpperCAmelCase : Tuple = TapasForMaskedLM(config=UpperCamelCase ) elif task == "INTERMEDIATE_PRETRAINING": __UpperCAmelCase : List[str] = TapasModel(config=UpperCamelCase ) else: raise ValueError(f"Task {task} not supported." ) print(f"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # Save pytorch-model (weights and configuration) print(f"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(UpperCamelCase ) # Save tokenizer files print(f"Save tokenizer files to {pytorch_dump_path}" ) __UpperCAmelCase : str = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + "vocab.txt" , model_max_length=512 ) tokenizer.save_pretrained(UpperCamelCase ) print("Used relative position embeddings:" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase = 1000 ) -> int: """simple docstring""" __UpperCAmelCase : List[Any] = 2**power __UpperCAmelCase : str = str(UpperCamelCase ) __UpperCAmelCase : Optional[int] = list(UpperCamelCase ) __UpperCAmelCase : List[str] = 0 for i in list_num: sum_of_num += int(UpperCamelCase ) return sum_of_num if __name__ == "__main__": A = int(input("""Enter the power of 2: """).strip()) print("""2 ^ """, power, """ = """, 2**power) A = solution(power) print("""Sum of the digits is: """, result)
77
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = "cpu" , UpperCamelCase = None ) -> None: """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.load(UpperCamelCase , map_location=UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(UpperCamelCase , torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) __UpperCAmelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __UpperCAmelCase : str = src_path torch.save(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
77
1
"""simple docstring""" import math A = 10 A = 7 A = BALLS_PER_COLOUR * NUM_COLOURS def _UpperCamelCase ( UpperCamelCase = 20 ) -> str: """simple docstring""" __UpperCAmelCase : Optional[int] = math.comb(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[int] = math.comb(NUM_BALLS - BALLS_PER_COLOUR , UpperCamelCase ) __UpperCAmelCase : Any = NUM_COLOURS * (1 - missing_colour / total) return f"{result:.9f}" if __name__ == "__main__": print(solution(20))
77
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": A = pd.read_csv("""sample_data.csv""", header=None) A = df.shape[:1][0] # If you're using some other dataset input the target column A = df.iloc[:, 1:2] A = actual_data.values.reshape(len_data, 1) A = MinMaxScaler().fit_transform(actual_data) A = 10 A = 5 A = 20 A = len_data - periods * look_back A = actual_data[:division] A = actual_data[division - look_back :] A , A = [], [] A , A = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) A = np.array(train_x) A = np.array(test_x) A = np.array([list(i.ravel()) for i in train_y]) A = np.array([list(i.ravel()) for i in test_y]) A = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") A = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) A = model.predict(x_test)
77
1
"""simple docstring""" from collections import namedtuple A = namedtuple("""from_to""", """from_ to""") A = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1_000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": from_to(0.76455, 1.30795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.000236588, 4226.75), } def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'from_type' value: {from_type!r} Supported values are:\n" + ", ".join(UpperCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n" + ", ".join(UpperCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A = 250_004 A = 250_020 @require_sentencepiece @require_tokenizers class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = MBartTokenizer lowercase_ = MBartTokenizerFast lowercase_ = True lowercase_ = True def a_ ( self : str): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase : Any = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) tokenizer.save_pretrained(self.tmpdirname) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Dict = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer.tokenize("This is a test") self.assertListEqual(UpperCamelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase_) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __UpperCAmelCase : List[Any] = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __UpperCAmelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def a_ ( self : Dict): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __UpperCAmelCase : Dict = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})"): __UpperCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = tempfile.mkdtemp() __UpperCAmelCase : Optional[int] = tokenizer_r.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Any = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) __UpperCAmelCase : Any = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f) self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Tuple = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=True __UpperCAmelCase : Optional[int] = tempfile.mkdtemp() __UpperCAmelCase : Dict = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : int = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=False __UpperCAmelCase : Tuple = tempfile.mkdtemp() __UpperCAmelCase : int = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way __UpperCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : str = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) @require_torch @require_sentencepiece @require_tokenizers class a__ ( unittest.TestCase ): lowercase_ = "facebook/mbart-large-en-ro" lowercase_ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowercase_ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowercase_ = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def a_ ( cls : int): """simple docstring""" __UpperCAmelCase : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en_XX" , tgt_lang="ro_RO") __UpperCAmelCase : Union[str, Any] = 1 return cls def a_ ( self : List[Any]): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ar_AR"] , 250001) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["en_EN"] , 250004) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ro_RO"] , 250020) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" self.assertIn(UpperCamelCase_ , self.tokenizer.all_special_ids) __UpperCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] __UpperCAmelCase : Optional[Any] = self.tokenizer.decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase_) self.assertEqual(UpperCamelCase_ , UpperCamelCase_) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , UpperCamelCase_) __UpperCAmelCase : Tuple = 10 __UpperCAmelCase : List[Any] = self.tokenizer(UpperCamelCase_ , max_length=UpperCamelCase_ , truncation=UpperCamelCase_).input_ids[0] self.assertEqual(ids[-2] , 2) self.assertEqual(ids[-1] , UpperCamelCase_) self.assertEqual(len(UpperCamelCase_) , UpperCamelCase_) def a_ ( self : Any): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"]) , [250026, 250001]) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[str] = tempfile.mkdtemp() __UpperCAmelCase : Union[str, Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase_) __UpperCAmelCase : List[Any] = MBartTokenizer.from_pretrained(UpperCamelCase_) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCamelCase_) @require_torch def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , return_tensors="pt") __UpperCAmelCase : Dict = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=len(self.expected_src_tokens) , return_tensors="pt" , ) __UpperCAmelCase : Tuple = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_) self.assertEqual((2, 14) , batch.input_ids.shape) self.assertEqual((2, 14) , batch.attention_mask.shape) __UpperCAmelCase : List[str] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) self.assertEqual(2 , batch.decoder_input_ids[0, -1]) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , []) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE]) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : List[str] = self.tokenizer(self.src_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=3 , return_tensors="pt") __UpperCAmelCase : Any = self.tokenizer( text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=10 , return_tensors="pt") __UpperCAmelCase : int = targets["input_ids"] __UpperCAmelCase : Any = shift_tokens_right(UpperCamelCase_ , self.tokenizer.pad_token_id) self.assertEqual(batch.input_ids.shape[1] , 3) self.assertEqual(batch.decoder_input_ids.shape[1] , 10) @require_torch def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="ar_AR") self.assertEqual( nested_simplify(UpperCamelCase_) , { # A, test, EOS, en_XX "input_ids": [[62, 3034, 2, 250004]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 250001, } , )
77
1
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets A = datasets.logging.get_logger(__name__) A = """\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel's Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = \"{COMET}: A Neural Framework for {MT} Evaluation\", author = \"Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon\", booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\", month = nov, year = \"2020\", address = \"Online\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\", pages = \"2685--2702\", } """ A = """\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA's or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. """ A = """ COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric('comet') >>> # comet_metric = load_metric('comet', 'wmt20-comet-da') # you can also choose which model to use >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"] >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"] >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results[\"scores\"]]) [0.19, 0.92] """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def a_ ( self : Optional[int]): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://unbabel.github.io/COMET/html/index.html" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "sources": datasets.Value("string" , id="sequence"), "predictions": datasets.Value("string" , id="sequence"), "references": datasets.Value("string" , id="sequence"), }) , codebase_urls=["https://github.com/Unbabel/COMET"] , reference_urls=[ "https://github.com/Unbabel/COMET", "https://www.aclweb.org/anthology/2020.emnlp-main.213/", "http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6", ] , ) def a_ ( self : str , UpperCamelCase_ : List[Any]): """simple docstring""" if self.config_name == "default": __UpperCAmelCase : str = comet.load_from_checkpoint(comet.download_model("wmt20-comet-da")) else: __UpperCAmelCase : Dict = comet.load_from_checkpoint(comet.download_model(self.config_name)) def a_ ( self : List[Any] , UpperCamelCase_ : int , UpperCamelCase_ : int , UpperCamelCase_ : int , UpperCamelCase_ : List[str]=None , UpperCamelCase_ : Tuple=False): """simple docstring""" if gpus is None: __UpperCAmelCase : Optional[Any] = 1 if torch.cuda.is_available() else 0 __UpperCAmelCase : List[str] = {"src": sources, "mt": predictions, "ref": references} __UpperCAmelCase : str = [dict(zip(UpperCamelCase_ , UpperCamelCase_)) for t in zip(*data.values())] __UpperCAmelCase , __UpperCAmelCase : Optional[int] = self.scorer.predict(UpperCamelCase_ , gpus=UpperCamelCase_ , progress_bar=UpperCamelCase_) return {"mean_score": mean_score, "scores": scores}
77
"""simple docstring""" from typing import Any class a__ : def __init__( self : List[str] , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : str = data __UpperCAmelCase : Optional[Any] = None class a__ : def __init__( self : Any): """simple docstring""" __UpperCAmelCase : Optional[int] = None def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.head while temp is not None: print(temp.data , end=" ") __UpperCAmelCase : Tuple = temp.next print() def a_ ( self : int , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : List[str] = Node(UpperCamelCase_) __UpperCAmelCase : str = self.head __UpperCAmelCase : Optional[int] = new_node def a_ ( self : str , UpperCamelCase_ : List[Any] , UpperCamelCase_ : str): """simple docstring""" if node_data_a == node_data_a: return else: __UpperCAmelCase : int = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Tuple = node_a.next __UpperCAmelCase : List[Any] = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Optional[Any] = node_a.next if node_a is None or node_a is None: return __UpperCAmelCase , __UpperCAmelCase : Any = node_a.data, node_a.data if __name__ == "__main__": A = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : Tuple = len(UpperCamelCase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : Optional[Any] = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(UpperCamelCase ): return None __UpperCAmelCase : List[Any] = sorted_collection[point] if current_item == item: return point else: if point < left: __UpperCAmelCase : Optional[int] = left __UpperCAmelCase : Any = point elif point > right: __UpperCAmelCase : Any = right __UpperCAmelCase : Optional[Any] = point else: if item < current_item: __UpperCAmelCase : List[str] = point - 1 else: __UpperCAmelCase : Optional[int] = point + 1 return None def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __UpperCAmelCase : Tuple = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(UpperCamelCase ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) elif point > right: return interpolation_search_by_recursion(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( UpperCamelCase , UpperCamelCase , UpperCamelCase , point - 1 ) else: return interpolation_search_by_recursion( UpperCamelCase , UpperCamelCase , point + 1 , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" if collection != sorted(UpperCamelCase ): raise ValueError("Collection must be ascending sorted" ) return True if __name__ == "__main__": import sys A = 0 if debug == 1: A = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("""Sequence must be ascending sorted to apply interpolation search""") A = 67 A = interpolation_search(collection, target) if result is not None: print(f'''{target} found at positions: {result}''') else: print("""Not found""")
77
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore A = """ Human: <<task>> Assistant: """ A = """huggingface-tools/default-prompts""" A = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase="run" ) -> List[str]: """simple docstring""" if prompt_or_repo_id is None: __UpperCAmelCase : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , UpperCamelCase ) is not None: return prompt_or_repo_id __UpperCAmelCase : str = cached_file( UpperCamelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(UpperCamelCase , "r" , encoding="utf-8" ) as f: return f.read()
77
1
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A = { """configuration_autoformer""": [ """AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """AutoformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """AutoformerForPrediction""", """AutoformerModel""", """AutoformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available A = { """configuration_ernie""": ["""ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ErnieConfig""", """ErnieOnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST""", """ErnieForCausalLM""", """ErnieForMaskedLM""", """ErnieForMultipleChoice""", """ErnieForNextSentencePrediction""", """ErnieForPreTraining""", """ErnieForQuestionAnswering""", """ErnieForSequenceClassification""", """ErnieForTokenClassification""", """ErnieModel""", """ErniePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): A = """pt""" elif is_tf_available(): A = """tf""" else: A = """jax""" class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ByTaTokenizer lowercase_ = False def a_ ( self : Any): """simple docstring""" super().setUp() __UpperCAmelCase : List[str] = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname) @cached_property def a_ ( self : List[str]): """simple docstring""" return ByTaTokenizer.from_pretrained("google/byt5-small") def a_ ( self : Any , **UpperCamelCase_ : str): """simple docstring""" return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : Tuple , UpperCamelCase_ : List[str]=False , UpperCamelCase_ : Dict=20 , UpperCamelCase_ : Dict=5): """simple docstring""" __UpperCAmelCase : Optional[Any] = [] for i in range(len(UpperCamelCase_)): try: __UpperCAmelCase : Optional[int] = tokenizer.decode([i] , clean_up_tokenization_spaces=UpperCamelCase_) except UnicodeDecodeError: pass toks.append((i, tok)) __UpperCAmelCase : List[str] = list(filter(lambda UpperCamelCase_: re.match(r"^[ a-zA-Z]+$" , t[1]) , UpperCamelCase_)) __UpperCAmelCase : Dict = list(filter(lambda UpperCamelCase_: [t[0]] == tokenizer.encode(t[1] , add_special_tokens=UpperCamelCase_) , UpperCamelCase_)) if max_length is not None and len(UpperCamelCase_) > max_length: __UpperCAmelCase : Optional[int] = toks[:max_length] if min_length is not None and len(UpperCamelCase_) < min_length and len(UpperCamelCase_) > 0: while len(UpperCamelCase_) < min_length: __UpperCAmelCase : Dict = toks + toks # toks_str = [t[1] for t in toks] __UpperCAmelCase : Optional[int] = [t[0] for t in toks] # Ensure consistency __UpperCAmelCase : List[str] = tokenizer.decode(UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_) if " " not in output_txt and len(UpperCamelCase_) > 1: __UpperCAmelCase : Optional[int] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=UpperCamelCase_) + " " + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=UpperCamelCase_) ) if with_prefix_space: __UpperCAmelCase : Any = " " + output_txt __UpperCAmelCase : str = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_) return output_txt, output_ids def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Any = self.ta_base_tokenizer __UpperCAmelCase : Any = tokenizer(["hi</s>", "I went to the gym</s>", "</s>"]) __UpperCAmelCase : List[str] = tokenizer(["hi", "I went to the gym", ""]) self.assertListEqual(batch_with_eos_added["input_ids"] , batch_without_eos_added["input_ids"]) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : str = self.ta_base_tokenizer __UpperCAmelCase : Tuple = "Unicode €." __UpperCAmelCase : List[str] = tokenizer(UpperCamelCase_) __UpperCAmelCase : Optional[int] = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded["input_ids"] , UpperCamelCase_) # decoding __UpperCAmelCase : Optional[Any] = tokenizer.decode(UpperCamelCase_) self.assertEqual(UpperCamelCase_ , "Unicode €.</s>") __UpperCAmelCase : Any = tokenizer("e è é ê ë") __UpperCAmelCase : Any = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded["input_ids"] , UpperCamelCase_) # decoding __UpperCAmelCase : str = tokenizer.decode(UpperCamelCase_) self.assertEqual(UpperCamelCase_ , "e è é ê ë</s>") # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode("e è é ê ë")) , "e è é ê ë</s>") def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : str = self.ta_base_tokenizer __UpperCAmelCase : Optional[int] = ["A long paragraph for summarization.", "Another paragraph for summarization."] # fmt: off __UpperCAmelCase : List[Any] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on __UpperCAmelCase : Any = tokenizer(UpperCamelCase_ , padding=UpperCamelCase_ , return_tensors=UpperCamelCase_) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_) if FRAMEWORK != "jax": __UpperCAmelCase : str = list(batch.input_ids.numpy()[0]) else: __UpperCAmelCase : str = list(batch.input_ids.tolist()[0]) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_) self.assertEqual((2, 37) , batch.input_ids.shape) self.assertEqual((2, 37) , batch.attention_mask.shape) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Tuple = self.ta_base_tokenizer __UpperCAmelCase : Dict = ["A long paragraph for summarization.", "Another paragraph for summarization."] __UpperCAmelCase : Optional[Any] = tokenizer(UpperCamelCase_ , padding=UpperCamelCase_ , return_tensors=UpperCamelCase_) # check if input_ids are returned and no decoder_input_ids self.assertIn("input_ids" , UpperCamelCase_) self.assertIn("attention_mask" , UpperCamelCase_) self.assertNotIn("decoder_input_ids" , UpperCamelCase_) self.assertNotIn("decoder_attention_mask" , UpperCamelCase_) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Any = self.ta_base_tokenizer __UpperCAmelCase : Tuple = [ "Summary of the text.", "Another summary.", ] __UpperCAmelCase : Union[str, Any] = tokenizer( text_target=UpperCamelCase_ , max_length=32 , padding="max_length" , truncation=UpperCamelCase_ , return_tensors=UpperCamelCase_) self.assertEqual(32 , targets["input_ids"].shape[1]) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.ta_base_tokenizer __UpperCAmelCase : Any = ["A long paragraph for summarization. </s>"] __UpperCAmelCase : str = ["Summary of the text. </s>"] # fmt: off __UpperCAmelCase : Optional[Any] = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] __UpperCAmelCase : Dict = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on __UpperCAmelCase : List[str] = tokenizer(UpperCamelCase_ , text_target=UpperCamelCase_) self.assertEqual(UpperCamelCase_ , batch["input_ids"][0]) self.assertEqual(UpperCamelCase_ , batch["labels"][0]) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}"): self.assertNotEqual(tokenizer.model_max_length , 42) # Now let's start the test __UpperCAmelCase : str = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}"): # Isolate this from the other tests because we save additional tokens/etc __UpperCAmelCase : List[Any] = tempfile.mkdtemp() __UpperCAmelCase : int = " He is very happy, UNwant\u00E9d,running" __UpperCAmelCase : str = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_) tokenizer.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Tuple = tokenizer.__class__.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = after_tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_) shutil.rmtree(UpperCamelCase_) __UpperCAmelCase : Tuple = self.get_tokenizers(model_max_length=42) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}"): # Isolate this from the other tests because we save additional tokens/etc __UpperCAmelCase : Union[str, Any] = tempfile.mkdtemp() __UpperCAmelCase : Any = " He is very happy, UNwant\u00E9d,running" tokenizer.add_tokens(["bim", "bambam"]) __UpperCAmelCase : Optional[int] = tokenizer.additional_special_tokens additional_special_tokens.append("new_additional_special_token") tokenizer.add_special_tokens({"additional_special_tokens": additional_special_tokens}) __UpperCAmelCase : Optional[Any] = tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_) tokenizer.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = tokenizer.__class__.from_pretrained(UpperCamelCase_) __UpperCAmelCase : List[Any] = after_tokenizer.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_) self.assertIn("new_additional_special_token" , after_tokenizer.additional_special_tokens) self.assertEqual(after_tokenizer.model_max_length , 42) __UpperCAmelCase : List[Any] = tokenizer.__class__.from_pretrained(UpperCamelCase_ , model_max_length=43) self.assertEqual(tokenizer.model_max_length , 43) shutil.rmtree(UpperCamelCase_) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Tuple = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer())) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer())) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase_) with open(os.path.join(UpperCamelCase_ , "special_tokens_map.json") , encoding="utf-8") as json_file: __UpperCAmelCase : int = json.load(UpperCamelCase_) with open(os.path.join(UpperCamelCase_ , "tokenizer_config.json") , encoding="utf-8") as json_file: __UpperCAmelCase : Dict = json.load(UpperCamelCase_) __UpperCAmelCase : Dict = [F"<extra_id_{i}>" for i in range(125)] __UpperCAmelCase : Dict = added_tokens_extra_ids + [ "an_additional_special_token" ] __UpperCAmelCase : int = added_tokens_extra_ids + [ "an_additional_special_token" ] with open(os.path.join(UpperCamelCase_ , "special_tokens_map.json") , "w" , encoding="utf-8") as outfile: json.dump(UpperCamelCase_ , UpperCamelCase_) with open(os.path.join(UpperCamelCase_ , "tokenizer_config.json") , "w" , encoding="utf-8") as outfile: json.dump(UpperCamelCase_ , UpperCamelCase_) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files __UpperCAmelCase : List[Any] = tokenizer_class.from_pretrained( UpperCamelCase_ , ) self.assertIn( "an_additional_special_token" , tokenizer_without_change_in_init.additional_special_tokens) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ["an_additional_special_token"] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(["an_additional_special_token"])) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained __UpperCAmelCase : Any = added_tokens_extra_ids + [AddedToken("a_new_additional_special_token" , lstrip=UpperCamelCase_)] __UpperCAmelCase : Optional[Any] = tokenizer_class.from_pretrained( UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , ) self.assertIn("a_new_additional_special_token" , tokenizer.additional_special_tokens) self.assertEqual( ["a_new_additional_special_token"] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(["a_new_additional_special_token"])) , ) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : List[str] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer())) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer())) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Any = tokenizer_class.from_pretrained(UpperCamelCase_) self.assertTrue(tokenizer.decode([255]) == "") def a_ ( self : List[str]): """simple docstring""" pass def a_ ( self : List[Any]): """simple docstring""" pass def a_ ( self : Any): """simple docstring""" pass def a_ ( self : Tuple): """simple docstring""" pass def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.get_tokenizers(fast=UpperCamelCase_ , do_lower_case=UpperCamelCase_) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}"): __UpperCAmelCase : Optional[Any] = ["t", "h", "i", "s", " ", "i", "s", " ", "a", " ", "t", "e", "x", "t", "</s>"] __UpperCAmelCase : Optional[Any] = tokenizer.convert_tokens_to_string(UpperCamelCase_) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}"): __UpperCAmelCase : List[Any] = [ "bos_token", "eos_token", "unk_token", "sep_token", "pad_token", "cls_token", "mask_token", ] __UpperCAmelCase : Dict = 0 __UpperCAmelCase : str = tokenizer.convert_ids_to_tokens( UpperCamelCase_ , skip_special_tokens=UpperCamelCase_) for attr in attributes_list: setattr(UpperCamelCase_ , attr + "_id" , UpperCamelCase_) self.assertEqual(getattr(UpperCamelCase_ , UpperCamelCase_) , UpperCamelCase_) self.assertEqual(getattr(UpperCamelCase_ , attr + "_id") , UpperCamelCase_) setattr(UpperCamelCase_ , attr + "_id" , UpperCamelCase_) self.assertEqual(getattr(UpperCamelCase_ , UpperCamelCase_) , UpperCamelCase_) self.assertEqual(getattr(UpperCamelCase_ , attr + "_id") , UpperCamelCase_) setattr(UpperCamelCase_ , "additional_special_tokens_ids" , []) self.assertListEqual(getattr(UpperCamelCase_ , "additional_special_tokens") , []) self.assertListEqual(getattr(UpperCamelCase_ , "additional_special_tokens_ids") , []) setattr(UpperCamelCase_ , "additional_special_tokens_ids" , [token_id_to_test_setters]) self.assertListEqual(getattr(UpperCamelCase_ , "additional_special_tokens") , [token_to_test_setters]) self.assertListEqual(getattr(UpperCamelCase_ , "additional_special_tokens_ids") , [token_id_to_test_setters])
77
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class a__ ( nn.Module ): def __init__( self : Union[str, Any]): """simple docstring""" super().__init__() __UpperCAmelCase : Optional[int] = nn.Linear(3 , 4) __UpperCAmelCase : str = nn.BatchNormad(4) __UpperCAmelCase : int = nn.Linear(4 , 5) def a_ ( self : str , UpperCamelCase_ : List[str]): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCamelCase_))) class a__ ( unittest.TestCase ): def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Optional[Any] = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , model.state_dict()) __UpperCAmelCase : Union[str, Any] = os.path.join(UpperCamelCase_ , "index.json") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: __UpperCAmelCase : Optional[int] = os.path.join(UpperCamelCase_ , F"{key}.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on the fact weights are properly loaded def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : int = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: __UpperCAmelCase : List[Any] = torch.randn(2 , 3 , dtype=UpperCamelCase_) with TemporaryDirectory() as tmp_dir: __UpperCAmelCase : Tuple = offload_weight(UpperCamelCase_ , "weight" , UpperCamelCase_ , {}) __UpperCAmelCase : Dict = os.path.join(UpperCamelCase_ , "weight.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) self.assertDictEqual(UpperCamelCase_ , {"weight": {"shape": [2, 3], "dtype": str(UpperCamelCase_).split(".")[1]}}) __UpperCAmelCase : Optional[Any] = load_offloaded_weight(UpperCamelCase_ , index["weight"]) self.assertTrue(torch.equal(UpperCamelCase_ , UpperCamelCase_)) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : List[Any] = ModelForTest() __UpperCAmelCase : Optional[int] = model.state_dict() __UpperCAmelCase : List[str] = {k: v for k, v in state_dict.items() if "linear2" not in k} __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "linear2" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : List[str] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "weight" in k} __UpperCAmelCase : Optional[Any] = {k: v for k, v in state_dict.items() if "weight" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[Any] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) # Duplicates are removed __UpperCAmelCase : str = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Any = {"a.1": 0, "a.10": 1, "a.2": 2} __UpperCAmelCase : Union[str, Any] = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1": 0, "a.2": 2}) __UpperCAmelCase : int = {"a.1.a": 0, "a.10.a": 1, "a.2.a": 2} __UpperCAmelCase : int = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1.a": 0, "a.2.a": 2})
77
1
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: """simple docstring""" # Construct model if openai_config_file == "": __UpperCAmelCase : List[str] = OpenAIGPTConfig() else: __UpperCAmelCase : Dict = OpenAIGPTConfig.from_json_file(UpperCamelCase ) __UpperCAmelCase : Tuple = OpenAIGPTModel(UpperCamelCase ) # Load weights from numpy load_tf_weights_in_openai_gpt(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # Save pytorch-model __UpperCAmelCase : Union[str, Any] = pytorch_dump_folder_path + "/" + WEIGHTS_NAME __UpperCAmelCase : Dict = pytorch_dump_folder_path + "/" + CONFIG_NAME print(f"Save PyTorch model to {pytorch_weights_dump_path}" ) torch.save(model.state_dict() , UpperCamelCase ) print(f"Save configuration file to {pytorch_config_dump_path}" ) with open(UpperCamelCase , "w" , encoding="utf-8" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) A = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
77
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Dict = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __UpperCAmelCase : Union[str, Any] = n - k # Calculate C(n,k) for i in range(UpperCamelCase ): result *= n - i result //= i + 1 return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , UpperCamelCase ) // (node_count + 1) def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if n < 0: raise ValueError("factorial() not defined for negative values" ) __UpperCAmelCase : Optional[Any] = 1 for i in range(1 , n + 1 ): result *= i return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return catalan_number(UpperCamelCase ) * factorial(UpperCamelCase ) if __name__ == "__main__": A = int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
77
1
"""simple docstring""" from __future__ import annotations def _UpperCamelCase ( UpperCamelCase ) -> list[int]: """simple docstring""" if len(UpperCamelCase ) == 0: return array __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = min(UpperCamelCase ), max(UpperCamelCase ) # Compute the variables __UpperCAmelCase : List[Any] = _max - _min + 1 __UpperCAmelCase , __UpperCAmelCase : List[str] = [0] * holes_range, [0] * holes_range # Make the sorting. for i in array: __UpperCAmelCase : List[str] = i - _min __UpperCAmelCase : Optional[Any] = i holes_repeat[index] += 1 # Makes the array back by replacing the numbers. __UpperCAmelCase : str = 0 for i in range(UpperCamelCase ): while holes_repeat[i] > 0: __UpperCAmelCase : str = holes[i] index += 1 holes_repeat[i] -= 1 # Returns the sorted array. return array if __name__ == "__main__": import doctest doctest.testmod() A = input("""Enter numbers separated by comma:\n""") A = [int(x) for x in user_input.split(""",""")] print(pigeon_sort(unsorted))
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) A = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" import argparse from argparse import Namespace import torch from torch import nn from transformers import XGLMConfig, XGLMForCausalLM def _UpperCamelCase ( UpperCamelCase ) -> Tuple: """simple docstring""" __UpperCAmelCase : int = [ "decoder.version", "decoder.output_projection.weight", "_float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase ) -> List[Any]: """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Any = emb.weight.shape __UpperCAmelCase : Dict = nn.Linear(UpperCamelCase , UpperCamelCase , bias=UpperCamelCase ) __UpperCAmelCase : Optional[Any] = emb.weight.data return lin_layer def _UpperCamelCase ( UpperCamelCase ) -> Dict: """simple docstring""" __UpperCAmelCase : Tuple = torch.load(UpperCamelCase , map_location="cpu" ) __UpperCAmelCase : Tuple = Namespace(**checkpoint["cfg"]["model"] ) __UpperCAmelCase : str = checkpoint["model"] remove_ignore_keys_(UpperCamelCase ) __UpperCAmelCase : int = state_dict["decoder.embed_tokens.weight"].shape[0] __UpperCAmelCase : Optional[int] = {key.replace("decoder" , "model" ): val for key, val in state_dict.items()} __UpperCAmelCase : Any = XGLMConfig( vocab_size=UpperCamelCase , max_position_embeddings=args.max_target_positions , num_layers=args.decoder_layers , attention_heads=args.decoder_attention_heads , ffn_dim=args.decoder_ffn_embed_dim , d_model=args.decoder_embed_dim , layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="gelu" , scale_embedding=not args.no_scale_embedding , tie_word_embeddings=args.share_decoder_input_output_embed , ) __UpperCAmelCase : Union[str, Any] = XGLMForCausalLM(UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = model.load_state_dict(UpperCamelCase , strict=UpperCamelCase ) print(UpperCamelCase ) __UpperCAmelCase : int = make_linear_from_emb(model.model.embed_tokens ) return model if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") A = parser.parse_args() A = convert_fairseq_xglm_checkpoint_from_disk(args.fairseq_path) model.save_pretrained(args.pytorch_dump_folder_path)
77
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCamelCase_ : int , UpperCamelCase_ : Any=13 , UpperCamelCase_ : Optional[int]=3 , UpperCamelCase_ : int=224 , UpperCamelCase_ : int=30 , UpperCamelCase_ : str=400 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Optional[int]=[0.5, 0.5, 0.5] , UpperCamelCase_ : Optional[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __UpperCAmelCase : Tuple = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : List[Any] = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : List[Any] = image_size __UpperCAmelCase : str = min_resolution __UpperCAmelCase : Tuple = max_resolution __UpperCAmelCase : Optional[Any] = do_resize __UpperCAmelCase : Any = size __UpperCAmelCase : Any = do_normalize __UpperCAmelCase : Any = image_mean __UpperCAmelCase : Optional[Any] = image_std def a_ ( self : str): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ViTImageProcessor if is_vision_available() else None def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Optional[Any] = EfficientFormerImageProcessorTester(self) @property def a_ ( self : Union[str, Any]): """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCamelCase_ , "image_std")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) def a_ ( self : Dict): """simple docstring""" pass def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : str = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Optional[int] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : Tuple = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Any = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
77
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A = logging.get_logger(__name__) A = { """hustvl/yolos-small""": """https://huggingface.co/hustvl/yolos-small/resolve/main/config.json""", # See all YOLOS models at https://huggingface.co/models?filter=yolos } class a__ ( __magic_name__ ): lowercase_ = "yolos" def __init__( self : Tuple , UpperCamelCase_ : int=768 , UpperCamelCase_ : Optional[int]=12 , UpperCamelCase_ : str=12 , UpperCamelCase_ : Union[str, Any]=3072 , UpperCamelCase_ : Any="gelu" , UpperCamelCase_ : int=0.0 , UpperCamelCase_ : Any=0.0 , UpperCamelCase_ : Optional[int]=0.02 , UpperCamelCase_ : List[str]=1e-12 , UpperCamelCase_ : Optional[int]=[512, 864] , UpperCamelCase_ : str=16 , UpperCamelCase_ : str=3 , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : str=100 , UpperCamelCase_ : str=True , UpperCamelCase_ : Union[str, Any]=False , UpperCamelCase_ : Optional[int]=1 , UpperCamelCase_ : Any=5 , UpperCamelCase_ : Tuple=2 , UpperCamelCase_ : List[Any]=5 , UpperCamelCase_ : List[Any]=2 , UpperCamelCase_ : Optional[Any]=0.1 , **UpperCamelCase_ : Any , ): """simple docstring""" super().__init__(**UpperCamelCase_) __UpperCAmelCase : Tuple = hidden_size __UpperCAmelCase : List[str] = num_hidden_layers __UpperCAmelCase : Tuple = num_attention_heads __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : int = hidden_act __UpperCAmelCase : int = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : List[Any] = initializer_range __UpperCAmelCase : int = layer_norm_eps __UpperCAmelCase : List[str] = image_size __UpperCAmelCase : Tuple = patch_size __UpperCAmelCase : int = num_channels __UpperCAmelCase : Optional[Any] = qkv_bias __UpperCAmelCase : List[Any] = num_detection_tokens __UpperCAmelCase : Union[str, Any] = use_mid_position_embeddings __UpperCAmelCase : List[str] = auxiliary_loss # Hungarian matcher __UpperCAmelCase : Optional[Any] = class_cost __UpperCAmelCase : Union[str, Any] = bbox_cost __UpperCAmelCase : int = giou_cost # Loss coefficients __UpperCAmelCase : Optional[int] = bbox_loss_coefficient __UpperCAmelCase : Union[str, Any] = giou_loss_coefficient __UpperCAmelCase : Any = eos_coefficient class a__ ( __magic_name__ ): lowercase_ = version.parse("1.11" ) @property def a_ ( self : int): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ]) @property def a_ ( self : List[Any]): """simple docstring""" return 1e-4 @property def a_ ( self : Optional[int]): """simple docstring""" return 12
77
"""simple docstring""" from collections import namedtuple A = namedtuple("""from_to""", """from_ to""") A = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1_000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": from_to(0.76455, 1.30795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.000236588, 4226.75), } def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'from_type' value: {from_type!r} Supported values are:\n" + ", ".join(UpperCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n" + ", ".join(UpperCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if not grid or not grid[0]: raise TypeError("The grid does not contain the appropriate information" ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] __UpperCAmelCase : List[Any] = grid[0] for row_n in range(1 , len(UpperCamelCase ) ): __UpperCAmelCase : Optional[int] = grid[row_n] __UpperCAmelCase : List[str] = fill_row(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : List[Any] = grid[row_n] return grid[-1][-1] def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> list: """simple docstring""" current_row[0] += row_above[0] for cell_n in range(1 , len(UpperCamelCase ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ShapEPipeline lowercase_ = ["prompt"] lowercase_ = ["prompt"] lowercase_ = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] lowercase_ = False @property def a_ ( self : Optional[int]): """simple docstring""" return 32 @property def a_ ( self : Any): """simple docstring""" return 32 @property def a_ ( self : int): """simple docstring""" return self.time_input_dim * 4 @property def a_ ( self : List[Any]): """simple docstring""" return 8 @property def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") return tokenizer @property def a_ ( self : List[str]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(UpperCamelCase_) @property def a_ ( self : Any): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Union[str, Any] = { "num_attention_heads": 2, "attention_head_dim": 16, "embedding_dim": self.time_input_dim, "num_embeddings": 32, "embedding_proj_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "num_layers": 1, "clip_embed_dim": self.time_input_dim * 2, "additional_embeddings": 0, "time_embed_act_fn": "gelu", "norm_in_type": "layer", "encoder_hid_proj_type": None, "added_emb_type": None, } __UpperCAmelCase : Dict = PriorTransformer(**UpperCamelCase_) return model @property def a_ ( self : Union[str, Any]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Tuple = { "param_shapes": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), "d_latent": self.time_input_dim, "d_hidden": self.renderer_dim, "n_output": 12, "background": ( 0.1, 0.1, 0.1, ), } __UpperCAmelCase : List[Any] = ShapERenderer(**UpperCamelCase_) return model def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.dummy_prior __UpperCAmelCase : str = self.dummy_text_encoder __UpperCAmelCase : int = self.dummy_tokenizer __UpperCAmelCase : int = self.dummy_renderer __UpperCAmelCase : Tuple = HeunDiscreteScheduler( beta_schedule="exp" , num_train_timesteps=1024 , prediction_type="sample" , use_karras_sigmas=UpperCamelCase_ , clip_sample=UpperCamelCase_ , clip_sample_range=1.0 , ) __UpperCAmelCase : str = { "prior": prior, "text_encoder": text_encoder, "tokenizer": tokenizer, "renderer": renderer, "scheduler": scheduler, } return components def a_ ( self : Optional[int] , UpperCamelCase_ : Dict , UpperCamelCase_ : Any=0): """simple docstring""" if str(UpperCamelCase_).startswith("mps"): __UpperCAmelCase : List[Any] = torch.manual_seed(UpperCamelCase_) else: __UpperCAmelCase : str = torch.Generator(device=UpperCamelCase_).manual_seed(UpperCamelCase_) __UpperCAmelCase : List[Any] = { "prompt": "horse", "generator": generator, "num_inference_steps": 1, "frame_size": 32, "output_type": "np", } return inputs def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : str = "cpu" __UpperCAmelCase : Union[str, Any] = self.get_dummy_components() __UpperCAmelCase : Union[str, Any] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[Any] = pipe(**self.get_dummy_inputs(UpperCamelCase_)) __UpperCAmelCase : Union[str, Any] = output.images[0] __UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __UpperCAmelCase : Union[str, Any] = np.array( [ 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def a_ ( self : Tuple): """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2]) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch_device == "cpu" __UpperCAmelCase : Optional[Any] = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=UpperCamelCase_ , relax_max_difference=UpperCamelCase_ , ) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.get_dummy_components() __UpperCAmelCase : List[str] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : int = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[int] = 1 __UpperCAmelCase : Any = 2 __UpperCAmelCase : Optional[Any] = self.get_dummy_inputs(UpperCamelCase_) for key in inputs.keys(): if key in self.batch_params: __UpperCAmelCase : List[Any] = batch_size * [inputs[key]] __UpperCAmelCase : List[Any] = pipe(**UpperCamelCase_ , num_images_per_prompt=UpperCamelCase_)[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class a__ ( unittest.TestCase ): def a_ ( self : List[str]): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/test_shap_e_np_out.npy") __UpperCAmelCase : Optional[Any] = ShapEPipeline.from_pretrained("openai/shap-e") __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Dict = torch.Generator(device=UpperCamelCase_).manual_seed(0) __UpperCAmelCase : int = pipe( "a shark" , generator=UpperCamelCase_ , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type="np" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(UpperCamelCase_ , UpperCamelCase_)
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { """bigcode/gpt_bigcode-santacoder""": """https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json""", } class a__ ( __magic_name__ ): lowercase_ = "gpt_bigcode" lowercase_ = ["past_key_values"] lowercase_ = { "hidden_size": "n_embd", "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Dict , UpperCamelCase_ : Tuple=50257 , UpperCamelCase_ : Union[str, Any]=1024 , UpperCamelCase_ : Any=768 , UpperCamelCase_ : Union[str, Any]=12 , UpperCamelCase_ : Optional[int]=12 , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Any="gelu_pytorch_tanh" , UpperCamelCase_ : Optional[int]=0.1 , UpperCamelCase_ : Union[str, Any]=0.1 , UpperCamelCase_ : List[str]=0.1 , UpperCamelCase_ : Dict=1e-5 , UpperCamelCase_ : Dict=0.02 , UpperCamelCase_ : List[str]=True , UpperCamelCase_ : List[str]=True , UpperCamelCase_ : Optional[int]=50256 , UpperCamelCase_ : Optional[int]=50256 , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Union[str, Any]=True , UpperCamelCase_ : Union[str, Any]=True , **UpperCamelCase_ : List[Any] , ): """simple docstring""" __UpperCAmelCase : List[str] = vocab_size __UpperCAmelCase : Union[str, Any] = n_positions __UpperCAmelCase : int = n_embd __UpperCAmelCase : Tuple = n_layer __UpperCAmelCase : str = n_head __UpperCAmelCase : List[Any] = n_inner __UpperCAmelCase : int = activation_function __UpperCAmelCase : Optional[int] = resid_pdrop __UpperCAmelCase : Union[str, Any] = embd_pdrop __UpperCAmelCase : Union[str, Any] = attn_pdrop __UpperCAmelCase : List[Any] = layer_norm_epsilon __UpperCAmelCase : Optional[Any] = initializer_range __UpperCAmelCase : Optional[Any] = scale_attn_weights __UpperCAmelCase : int = use_cache __UpperCAmelCase : int = attention_softmax_in_fpaa __UpperCAmelCase : Optional[Any] = scale_attention_softmax_in_fpaa __UpperCAmelCase : Union[str, Any] = multi_query __UpperCAmelCase : str = bos_token_id __UpperCAmelCase : Union[str, Any] = eos_token_id super().__init__(bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_)
77
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_features", "is_longer"] def __init__( self : List[str] , UpperCamelCase_ : Dict=64 , UpperCamelCase_ : Tuple=48000 , UpperCamelCase_ : List[Any]=480 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=1024 , UpperCamelCase_ : List[str]=0.0 , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : float = 0 , UpperCamelCase_ : float = 14000 , UpperCamelCase_ : int = None , UpperCamelCase_ : str = "fusion" , UpperCamelCase_ : str = "repeatpad" , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" super().__init__( feature_size=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , padding_value=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = top_db __UpperCAmelCase : Optional[Any] = truncation __UpperCAmelCase : str = padding __UpperCAmelCase : int = fft_window_size __UpperCAmelCase : str = (fft_window_size >> 1) + 1 __UpperCAmelCase : List[Any] = hop_length __UpperCAmelCase : Optional[Any] = max_length_s __UpperCAmelCase : Tuple = max_length_s * sampling_rate __UpperCAmelCase : str = sampling_rate __UpperCAmelCase : int = frequency_min __UpperCAmelCase : Optional[Any] = frequency_max __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm=UpperCamelCase_ , mel_scale="htk" , ) __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm="slaney" , mel_scale="slaney" , ) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Dict = copy.deepcopy(self.__dict__) __UpperCAmelCase : str = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def a_ ( self : int , UpperCamelCase_ : np.array , UpperCamelCase_ : Optional[np.array] = None): """simple docstring""" __UpperCAmelCase : List[Any] = spectrogram( UpperCamelCase_ , window_function(self.fft_window_size , "hann") , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=UpperCamelCase_ , log_mel="dB" , ) return log_mel_spectrogram.T def a_ ( self : List[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = np.array_split(list(range(0 , total_frames - chunk_frames + 1)) , 3) if len(ranges[1]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : str = [0] if len(ranges[2]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : Dict = [0] # randomly choose index for each part __UpperCAmelCase : Dict = np.random.choice(ranges[0]) __UpperCAmelCase : List[str] = np.random.choice(ranges[1]) __UpperCAmelCase : List[Any] = np.random.choice(ranges[2]) __UpperCAmelCase : List[Any] = mel[idx_front : idx_front + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_back : idx_back + chunk_frames, :] __UpperCAmelCase : Tuple = torch.tensor(mel[None, None, :]) __UpperCAmelCase : Union[str, Any] = torch.nn.functional.interpolate( UpperCamelCase_ , size=[chunk_frames, 64] , mode="bilinear" , align_corners=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = mel_shrink[0][0].numpy() __UpperCAmelCase : Optional[int] = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0) return mel_fusion def a_ ( self : Optional[Any] , UpperCamelCase_ : np.array , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[Any]): """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": __UpperCAmelCase : List[str] = True # random crop to max_length (for compatibility) -> this should be handled by self.pad __UpperCAmelCase : List[Any] = len(UpperCamelCase_) - max_length __UpperCAmelCase : int = np.random.randint(0 , overflow + 1) __UpperCAmelCase : Union[str, Any] = waveform[idx : idx + max_length] __UpperCAmelCase : Union[str, Any] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] elif truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : Dict = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed __UpperCAmelCase : Tuple = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. __UpperCAmelCase : List[str] = np.stack([mel, mel, mel, mel] , axis=0) __UpperCAmelCase : Any = False else: __UpperCAmelCase : List[str] = self._random_mel_fusion(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented") else: __UpperCAmelCase : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": __UpperCAmelCase : Tuple = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : List[str] = np.stack(np.tile(UpperCamelCase_ , n_repeat + 1))[:max_length] if padding == "repeatpad": __UpperCAmelCase : Union[str, Any] = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : Optional[Any] = np.stack(np.tile(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : int = np.pad(UpperCamelCase_ , (0, max_length - waveform.shape[0]) , mode="constant" , constant_values=0) if truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0) else: __UpperCAmelCase : Optional[int] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] return input_mel, longer def __call__( self : Dict , UpperCamelCase_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase_ : str = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[Union[str, TensorType]] = None , **UpperCamelCase_ : Any , ): """simple docstring""" __UpperCAmelCase : int = truncation if truncation is not None else self.truncation __UpperCAmelCase : Optional[Any] = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}.") else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug.") __UpperCAmelCase : List[str] = isinstance(UpperCamelCase_ , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}") __UpperCAmelCase : str = is_batched_numpy or ( isinstance(UpperCamelCase_ , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: __UpperCAmelCase : Dict = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase_ , np.ndarray): __UpperCAmelCase : Tuple = np.asarray(UpperCamelCase_ , dtype=np.floataa) elif isinstance(UpperCamelCase_ , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): __UpperCAmelCase : Optional[int] = raw_speech.astype(np.floataa) # always return batch if not is_batched: __UpperCAmelCase : int = [np.asarray(UpperCamelCase_)] # convert to mel spectrogram, truncate and pad if needed. __UpperCAmelCase : Optional[int] = [ self._get_input_mel(UpperCamelCase_ , max_length if max_length else self.nb_max_samples , UpperCamelCase_ , UpperCamelCase_) for waveform in raw_speech ] __UpperCAmelCase : Tuple = [] __UpperCAmelCase : List[Any] = [] for mel, longer in padded_inputs: input_mel.append(UpperCamelCase_) is_longer.append(UpperCamelCase_) if truncation == "fusion" and sum(UpperCamelCase_) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer __UpperCAmelCase : Any = np.random.randint(0 , len(UpperCamelCase_)) __UpperCAmelCase : Optional[int] = True if isinstance(input_mel[0] , UpperCamelCase_): __UpperCAmelCase : Tuple = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for feature in input_mel] # is_longer is a list of bool __UpperCAmelCase : List[str] = [[longer] for longer in is_longer] __UpperCAmelCase : Optional[int] = {"input_features": input_mel, "is_longer": is_longer} __UpperCAmelCase : Optional[int] = BatchFeature(UpperCamelCase_) if return_tensors is not None: __UpperCAmelCase : Any = input_features.convert_to_tensors(UpperCamelCase_) return input_features
77
1
"""simple docstring""" import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = None lowercase_ = BloomTokenizerFast lowercase_ = BloomTokenizerFast lowercase_ = True lowercase_ = False lowercase_ = "tokenizer_file" lowercase_ = {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def a_ ( self : Optional[Any]): """simple docstring""" super().setUp() __UpperCAmelCase : Tuple = BloomTokenizerFast.from_pretrained("bigscience/tokenizer") tokenizer.save_pretrained(self.tmpdirname) def a_ ( self : Optional[Any] , **UpperCamelCase_ : List[Any]): """simple docstring""" kwargs.update(self.special_tokens_map) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **UpperCamelCase_) def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : int = self.get_rust_tokenizer() __UpperCAmelCase : List[str] = ["The quick brown fox</s>", "jumps over the lazy dog</s>"] __UpperCAmelCase : Union[str, Any] = [[2175, 23714, 73173, 144252, 2], [77, 132619, 3478, 368, 109586, 35433, 2]] __UpperCAmelCase : List[Any] = tokenizer.batch_encode_plus(UpperCamelCase_)["input_ids"] self.assertListEqual(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : str = tokenizer.batch_decode(UpperCamelCase_) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : Tuple=6): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})"): __UpperCAmelCase : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input __UpperCAmelCase : Any = "This is a simple input" __UpperCAmelCase : str = ["This is a simple input 1", "This is a simple input 2"] __UpperCAmelCase : Union[str, Any] = ("This is a simple input", "This is a pair") __UpperCAmelCase : Union[str, Any] = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests try: tokenizer_r.encode(UpperCamelCase_ , max_length=UpperCamelCase_) tokenizer_r.encode_plus(UpperCamelCase_ , max_length=UpperCamelCase_) tokenizer_r.batch_encode_plus(UpperCamelCase_ , max_length=UpperCamelCase_) tokenizer_r.encode(UpperCamelCase_ , max_length=UpperCamelCase_) tokenizer_r.batch_encode_plus(UpperCamelCase_ , max_length=UpperCamelCase_) except ValueError: self.fail("Bloom Tokenizer should be able to deal with padding") __UpperCAmelCase : Tuple = None # Hotfixing padding = None self.assertRaises(UpperCamelCase_ , tokenizer_r.encode , UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length") # Simple input self.assertRaises(UpperCamelCase_ , tokenizer_r.encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length") # Simple input self.assertRaises( UpperCamelCase_ , tokenizer_r.batch_encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length" , ) # Pair input self.assertRaises(UpperCamelCase_ , tokenizer_r.encode , UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length") # Pair input self.assertRaises(UpperCamelCase_ , tokenizer_r.encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length") # Pair input self.assertRaises( UpperCamelCase_ , tokenizer_r.batch_encode_plus , UpperCamelCase_ , max_length=UpperCamelCase_ , padding="max_length" , ) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : int = self.get_rust_tokenizer() __UpperCAmelCase : Union[str, Any] = load_dataset("xnli" , "all_languages" , split="test" , streaming=UpperCamelCase_) __UpperCAmelCase : Any = next(iter(UpperCamelCase_))["premise"] # pick up one data __UpperCAmelCase : Optional[int] = list(sample_data.values()) __UpperCAmelCase : str = list(map(tokenizer.encode , UpperCamelCase_)) __UpperCAmelCase : Dict = [tokenizer.decode(UpperCamelCase_ , clean_up_tokenization_spaces=UpperCamelCase_) for x in output_tokens] self.assertListEqual(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : str): """simple docstring""" self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map) , 1) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values())[0]) , 1)
77
"""simple docstring""" import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , UpperCamelCase_ : List[Any]="</s>" , UpperCamelCase_ : Tuple="<unk>" , UpperCamelCase_ : List[str]="<pad>" , UpperCamelCase_ : Union[str, Any]=125 , UpperCamelCase_ : Dict=None , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: __UpperCAmelCase : int = [F"<extra_id_{i}>" for i in range(UpperCamelCase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens __UpperCAmelCase : Dict = len(set(filter(lambda UpperCamelCase_: bool("extra_id" in str(UpperCamelCase_)) , UpperCamelCase_))) if extra_tokens != extra_ids: raise ValueError( F"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" " provided to ByT5Tokenizer. In this case the additional_special_tokens must include the" " extra_ids tokens") __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else pad_token __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else eos_token __UpperCAmelCase : Optional[int] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else unk_token super().__init__( eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , extra_ids=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : List[str] = extra_ids __UpperCAmelCase : int = 2**8 # utf is 8 bits # define special tokens dict __UpperCAmelCase : Dict[int, str] = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } __UpperCAmelCase : Any = len(self.special_tokens_encoder) __UpperCAmelCase : List[Any] = len(UpperCamelCase_) for i, token in enumerate(UpperCamelCase_): __UpperCAmelCase : Union[str, Any] = self.vocab_size + i - n __UpperCAmelCase : Dict[str, int] = {v: k for k, v in self.special_tokens_encoder.items()} @property def a_ ( self : List[Any]): """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def a_ ( self : List[str] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(UpperCamelCase_)) + [1] return ([0] * len(UpperCamelCase_)) + [1] + ([0] * len(UpperCamelCase_)) + [1] def a_ ( self : Optional[Any] , UpperCamelCase_ : List[int]): """simple docstring""" if len(UpperCamelCase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F"This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated" " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def a_ ( self : Dict , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Dict = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def a_ ( self : Optional[int] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Optional[Any] = self._add_eos_if_not_present(UpperCamelCase_) if token_ids_a is None: return token_ids_a else: __UpperCAmelCase : List[Any] = self._add_eos_if_not_present(UpperCamelCase_) return token_ids_a + token_ids_a def a_ ( self : List[str] , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : Any = [chr(UpperCamelCase_) for i in text.encode("utf-8")] return tokens def a_ ( self : Tuple , UpperCamelCase_ : List[Any]): """simple docstring""" if token in self.special_tokens_encoder: __UpperCAmelCase : Any = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: __UpperCAmelCase : int = self.added_tokens_encoder[token] elif len(UpperCamelCase_) != 1: __UpperCAmelCase : Optional[Any] = self.unk_token_id else: __UpperCAmelCase : Any = ord(UpperCamelCase_) + self._num_special_tokens return token_id def a_ ( self : Any , UpperCamelCase_ : List[str]): """simple docstring""" if index in self.special_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[index] else: __UpperCAmelCase : List[str] = chr(index - self._num_special_tokens) return token def a_ ( self : Dict , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : str = b"" for token in tokens: if token in self.special_tokens_decoder: __UpperCAmelCase : Tuple = self.special_tokens_decoder[token].encode("utf-8") elif token in self.added_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[token].encode("utf-8") elif token in self.special_tokens_encoder: __UpperCAmelCase : Optional[int] = token.encode("utf-8") elif token in self.added_tokens_encoder: __UpperCAmelCase : Optional[Any] = token.encode("utf-8") else: __UpperCAmelCase : Any = bytes([ord(UpperCamelCase_)]) bstring += tok_string __UpperCAmelCase : List[Any] = bstring.decode("utf-8" , errors="ignore") return string def a_ ( self : Optional[int] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" return ()
77
1
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch A = random.Random() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=1.0 , UpperCamelCase=None , UpperCamelCase=None ) -> Optional[Any]: """simple docstring""" if rng is None: __UpperCAmelCase : Tuple = global_rng __UpperCAmelCase : str = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class a__ ( unittest.TestCase ): def __init__( self : Tuple , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Any=7 , UpperCamelCase_ : int=400 , UpperCamelCase_ : List[Any]=2000 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : List[Any]=160 , UpperCamelCase_ : str=8 , UpperCamelCase_ : str=0.0 , UpperCamelCase_ : Union[str, Any]=4000 , UpperCamelCase_ : Any=False , UpperCamelCase_ : List[str]=True , ): """simple docstring""" __UpperCAmelCase : Optional[Any] = parent __UpperCAmelCase : Dict = batch_size __UpperCAmelCase : List[str] = min_seq_length __UpperCAmelCase : Any = max_seq_length __UpperCAmelCase : Dict = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __UpperCAmelCase : Dict = padding_value __UpperCAmelCase : Any = sampling_rate __UpperCAmelCase : Any = return_attention_mask __UpperCAmelCase : Any = do_normalize __UpperCAmelCase : Optional[Any] = feature_size __UpperCAmelCase : Tuple = chunk_length __UpperCAmelCase : Any = hop_length def a_ ( self : Tuple): """simple docstring""" return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def a_ ( self : Optional[int] , UpperCamelCase_ : Tuple=False , UpperCamelCase_ : int=False): """simple docstring""" def _flatten(UpperCamelCase_ : Union[str, Any]): return list(itertools.chain(*UpperCamelCase_)) if equal_length: __UpperCAmelCase : Tuple = [floats_list((self.max_seq_length, self.feature_size)) for _ in range(self.batch_size)] else: # make sure that inputs increase in size __UpperCAmelCase : int = [ floats_list((x, self.feature_size)) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff) ] if numpify: __UpperCAmelCase : List[str] = [np.asarray(UpperCamelCase_) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = WhisperFeatureExtractor if is_speech_available() else None def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = WhisperFeatureExtractionTester(self) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Optional[int] = self.feature_extraction_class(**self.feat_extract_dict) with tempfile.TemporaryDirectory() as tmpdirname: __UpperCAmelCase : str = feat_extract_first.save_pretrained(UpperCamelCase_)[0] check_json_file_has_correct_format(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = self.feature_extraction_class.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Tuple = feat_extract_first.to_dict() __UpperCAmelCase : Any = feat_extract_second.to_dict() __UpperCAmelCase : int = feat_extract_first.mel_filters __UpperCAmelCase : List[Any] = feat_extract_second.mel_filters self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_)) self.assertEqual(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = self.feature_extraction_class(**self.feat_extract_dict) with tempfile.TemporaryDirectory() as tmpdirname: __UpperCAmelCase : str = os.path.join(UpperCamelCase_ , "feat_extract.json") feat_extract_first.to_json_file(UpperCamelCase_) __UpperCAmelCase : List[Any] = self.feature_extraction_class.from_json_file(UpperCamelCase_) __UpperCAmelCase : str = feat_extract_first.to_dict() __UpperCAmelCase : List[Any] = feat_extract_second.to_dict() __UpperCAmelCase : Dict = feat_extract_first.mel_filters __UpperCAmelCase : List[Any] = feat_extract_second.mel_filters self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_)) self.assertEqual(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : List[Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) # create three inputs of length 800, 1000, and 1200 __UpperCAmelCase : str = [floats_list((1, x))[0] for x in range(800 , 1400 , 200)] __UpperCAmelCase : int = [np.asarray(UpperCamelCase_) for speech_input in speech_inputs] # Test feature size __UpperCAmelCase : Tuple = feature_extractor(UpperCamelCase_ , padding="max_length" , return_tensors="np").input_features self.assertTrue(input_features.ndim == 3) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size) # Test not batched input __UpperCAmelCase : str = feature_extractor(speech_inputs[0] , return_tensors="np").input_features __UpperCAmelCase : Dict = feature_extractor(np_speech_inputs[0] , return_tensors="np").input_features self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3)) # Test batched __UpperCAmelCase : List[str] = feature_extractor(UpperCamelCase_ , return_tensors="np").input_features __UpperCAmelCase : str = feature_extractor(UpperCamelCase_ , return_tensors="np").input_features for enc_seq_a, enc_seq_a in zip(UpperCamelCase_ , UpperCamelCase_): self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3)) # Test 2-D numpy arrays are batched. __UpperCAmelCase : List[Any] = [floats_list((1, x))[0] for x in (800, 800, 800)] __UpperCAmelCase : Optional[int] = np.asarray(UpperCamelCase_) __UpperCAmelCase : List[Any] = feature_extractor(UpperCamelCase_ , return_tensors="np").input_features __UpperCAmelCase : Any = feature_extractor(UpperCamelCase_ , return_tensors="np").input_features for enc_seq_a, enc_seq_a in zip(UpperCamelCase_ , UpperCamelCase_): self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3)) # Test truncation required __UpperCAmelCase : Tuple = [floats_list((1, x))[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200)] __UpperCAmelCase : List[Any] = [np.asarray(UpperCamelCase_) for speech_input in speech_inputs] __UpperCAmelCase : Tuple = [x[: feature_extractor.n_samples] for x in speech_inputs] __UpperCAmelCase : int = [np.asarray(UpperCamelCase_) for speech_input in speech_inputs_truncated] __UpperCAmelCase : Optional[int] = feature_extractor(UpperCamelCase_ , return_tensors="np").input_features __UpperCAmelCase : Optional[int] = feature_extractor(UpperCamelCase_ , return_tensors="np").input_features for enc_seq_a, enc_seq_a in zip(UpperCamelCase_ , UpperCamelCase_): self.assertTrue(np.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3)) def a_ ( self : Union[str, Any]): """simple docstring""" import torch __UpperCAmelCase : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) __UpperCAmelCase : Dict = np.random.rand(100 , 32).astype(np.floataa) __UpperCAmelCase : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __UpperCAmelCase : str = feature_extractor.pad([{"input_features": inputs}] , return_tensors="np") self.assertTrue(np_processed.input_features.dtype == np.floataa) __UpperCAmelCase : Optional[Any] = feature_extractor.pad([{"input_features": inputs}] , return_tensors="pt") self.assertTrue(pt_processed.input_features.dtype == torch.floataa) def a_ ( self : Optional[Any] , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : Union[str, Any] = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation") # automatic decoding with librispeech __UpperCAmelCase : List[str] = ds.sort("id").select(range(UpperCamelCase_))[:num_samples]["audio"] return [x["array"] for x in speech_samples] def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Any = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ]) # fmt: on __UpperCAmelCase : Union[str, Any] = self._load_datasamples(1) __UpperCAmelCase : Optional[int] = WhisperFeatureExtractor() __UpperCAmelCase : Dict = feature_extractor(UpperCamelCase_ , return_tensors="pt").input_features self.assertEqual(input_features.shape , (1, 80, 3000)) self.assertTrue(torch.allclose(input_features[0, 0, :30] , UpperCamelCase_ , atol=1e-4)) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict()) __UpperCAmelCase : List[Any] = self._load_datasamples(1)[0] __UpperCAmelCase : Dict = ((audio - audio.min()) / (audio.max() - audio.min())) * 65535 # Rescale to [0, 65535] to show issue __UpperCAmelCase : Optional[Any] = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=UpperCamelCase_)[0] self.assertTrue(np.all(np.mean(UpperCamelCase_) < 1e-3)) self.assertTrue(np.all(np.abs(np.var(UpperCamelCase_) - 1) < 1e-3))
77
"""simple docstring""" import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a__ ( unittest.TestCase ): def __init__( self : List[Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple=3 , UpperCamelCase_ : Optional[int]=32 , UpperCamelCase_ : Dict=3 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=[10, 20, 30, 40] , UpperCamelCase_ : Tuple=[1, 1, 2, 1] , UpperCamelCase_ : str=True , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Dict="relu" , UpperCamelCase_ : str=3 , UpperCamelCase_ : int=None , ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : List[str] = image_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : Union[str, Any] = embeddings_size __UpperCAmelCase : Dict = hidden_sizes __UpperCAmelCase : Dict = depths __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_labels __UpperCAmelCase : Optional[int] = hidden_act __UpperCAmelCase : str = num_labels __UpperCAmelCase : Optional[int] = scope __UpperCAmelCase : Dict = len(UpperCamelCase_) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __UpperCAmelCase : Dict = self.get_config() return config, pixel_values def a_ ( self : Dict): """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def a_ ( self : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[str] = FlaxRegNetModel(config=UpperCamelCase_) __UpperCAmelCase : Dict = model(UpperCamelCase_) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a_ ( self : Any , UpperCamelCase_ : Dict , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : List[Any] = self.num_labels __UpperCAmelCase : Tuple = FlaxRegNetForImageClassification(config=UpperCamelCase_) __UpperCAmelCase : str = model(UpperCamelCase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Any = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase : Tuple = config_and_inputs __UpperCAmelCase : Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () lowercase_ = False lowercase_ = False lowercase_ = False def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = FlaxRegNetModelTester(self) __UpperCAmelCase : str = ConfigTester(self , config_class=UpperCamelCase_ , has_text_modality=UpperCamelCase_) def a_ ( self : Dict): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a_ ( self : Tuple): """simple docstring""" return def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase_) @unittest.skip(reason="RegNet does not use inputs_embeds") def a_ ( self : Union[str, Any]): """simple docstring""" pass @unittest.skip(reason="RegNet does not support input and output embeddings") def a_ ( self : Optional[int]): """simple docstring""" pass def a_ ( self : str): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : int = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[int] = inspect.signature(model.__call__) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : Any = [*signature.parameters.keys()] __UpperCAmelCase : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase_) def a_ ( self : int): """simple docstring""" def check_hidden_states_output(UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Union[str, Any]): __UpperCAmelCase : Union[str, Any] = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = model(**self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCAmelCase : str = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase_) , expected_num_stages + 1) __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : List[str] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase : Optional[int] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): __UpperCAmelCase : List[Any] = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[int] = model_class(UpperCamelCase_) @jax.jit def model_jitted(UpperCamelCase_ : int , **UpperCamelCase_ : Optional[int]): return model(pixel_values=UpperCamelCase_ , **UpperCamelCase_) with self.subTest("JIT Enabled"): __UpperCAmelCase : Optional[Any] = model_jitted(**UpperCamelCase_).to_tuple() with self.subTest("JIT Disabled"): with jax.disable_jit(): __UpperCAmelCase : Dict = model_jitted(**UpperCamelCase_).to_tuple() self.assertEqual(len(UpperCamelCase_) , len(UpperCamelCase_)) for jitted_output, output in zip(UpperCamelCase_ , UpperCamelCase_): self.assertEqual(jitted_output.shape , output.shape) def _UpperCamelCase ( ) -> Any: """simple docstring""" __UpperCAmelCase : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_flax class a__ ( unittest.TestCase ): @cached_property def a_ ( self : Optional[int]): """simple docstring""" return AutoImageProcessor.from_pretrained("facebook/regnet-y-040") if is_vision_available() else None @slow def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Any = FlaxRegNetForImageClassification.from_pretrained("facebook/regnet-y-040") __UpperCAmelCase : Dict = self.default_image_processor __UpperCAmelCase : str = prepare_img() __UpperCAmelCase : int = image_processor(images=UpperCamelCase_ , return_tensors="np") __UpperCAmelCase : Dict = model(**UpperCamelCase_) # verify the logits __UpperCAmelCase : Dict = (1, 1000) self.assertEqual(outputs.logits.shape , UpperCamelCase_) __UpperCAmelCase : Any = jnp.array([-0.4180, -1.5051, -3.4836]) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , UpperCamelCase_ , atol=1e-4))
77
1
"""simple docstring""" import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch A = True except ImportError: A = False try: from torch.hub import _get_torch_home A = _get_torch_home() except ImportError: A = os.path.expanduser( os.getenv("""TORCH_HOME""", os.path.join(os.getenv("""XDG_CACHE_HOME""", """~/.cache"""), """torch""")) ) A = os.path.join(torch_cache_home, """transformers""") A = """https://cdn.huggingface.co""" A = """https://s3.amazonaws.com/models.huggingface.co/bert""" A = """/""".join(str(Path(__file__).resolve()).split("""/""")[:-1]) A = os.path.join(PATH, """config.yaml""") A = os.path.join(PATH, """attributes.txt""") A = os.path.join(PATH, """objects.txt""") A = os.getenv("""PYTORCH_PRETRAINED_BERT_CACHE""", default_cache_path) A = os.getenv("""PYTORCH_TRANSFORMERS_CACHE""", PYTORCH_PRETRAINED_BERT_CACHE) A = os.getenv("""TRANSFORMERS_CACHE""", PYTORCH_TRANSFORMERS_CACHE) A = """pytorch_model.bin""" A = """config.yaml""" def _UpperCamelCase ( UpperCamelCase=OBJECTS , UpperCamelCase=ATTRIBUTES ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[Any] = [] with open(UpperCamelCase ) as f: for object in f.readlines(): vg_classes.append(object.split("," )[0].lower().strip() ) __UpperCAmelCase : Union[str, Any] = [] with open(UpperCamelCase ) as f: for object in f.readlines(): vg_attrs.append(object.split("," )[0].lower().strip() ) return vg_classes, vg_attrs def _UpperCamelCase ( UpperCamelCase ) -> List[str]: """simple docstring""" __UpperCAmelCase : str = OrderedDict() with open(UpperCamelCase , "rb" ) as f: __UpperCAmelCase : List[Any] = pkl.load(UpperCamelCase )["model"] for k in copy.deepcopy(list(ckp.keys() ) ): __UpperCAmelCase : Dict = ckp.pop(UpperCamelCase ) if isinstance(UpperCamelCase , np.ndarray ): __UpperCAmelCase : Any = torch.tensor(UpperCamelCase ) else: assert isinstance(UpperCamelCase , torch.tensor ), type(UpperCamelCase ) __UpperCAmelCase : Tuple = v return r class a__ : lowercase_ = {} def __init__( self : Any , UpperCamelCase_ : dict , UpperCamelCase_ : str = "root" , UpperCamelCase_ : List[str]=0): """simple docstring""" __UpperCAmelCase : Union[str, Any] = name __UpperCAmelCase : Any = level __UpperCAmelCase : Optional[Any] = {} for k, v in dictionary.items(): if v is None: raise ValueError() __UpperCAmelCase : str = copy.deepcopy(UpperCamelCase_) __UpperCAmelCase : Tuple = copy.deepcopy(UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_): __UpperCAmelCase : Dict = Config(UpperCamelCase_ , name=UpperCamelCase_ , level=level + 1) __UpperCAmelCase : str = v setattr(self , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Tuple = d def __repr__( self : str): """simple docstring""" return str(list((self._pointer.keys()))) def __setattr__( self : Tuple , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[int] = val __UpperCAmelCase : Optional[Any] = val __UpperCAmelCase : List[str] = key.split(".") __UpperCAmelCase : Union[str, Any] = len(UpperCamelCase_) - 1 __UpperCAmelCase : Any = self._pointer if len(UpperCamelCase_) > 1: for i, l in enumerate(UpperCamelCase_): if hasattr(self , UpperCamelCase_) and isinstance(getattr(self , UpperCamelCase_) , UpperCamelCase_): setattr(getattr(self , UpperCamelCase_) , ".".join(levels[i:]) , UpperCamelCase_) if l == last_level: __UpperCAmelCase : str = val else: __UpperCAmelCase : Optional[Any] = pointer[l] def a_ ( self : Union[str, Any]): """simple docstring""" return self._pointer def a_ ( self : str , UpperCamelCase_ : Dict , UpperCamelCase_ : Optional[int]): """simple docstring""" with open(F"{file_name}" , "w") as stream: dump(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Optional[int] , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Any): """simple docstring""" with open(F"{file_name}" , "w") as stream: json.dump(UpperCamelCase_ , UpperCamelCase_) @staticmethod def a_ ( UpperCamelCase_ : Any): """simple docstring""" with open(UpperCamelCase_) as stream: __UpperCAmelCase : int = load(UpperCamelCase_ , Loader=UpperCamelCase_) return data def __str__( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = " " if self._name != "root": __UpperCAmelCase : str = F"{t * (self._level-1)}{self._name}:\n" else: __UpperCAmelCase : str = "" __UpperCAmelCase : int = self._level for i, (k, v) in enumerate(self._pointer.items()): if isinstance(UpperCamelCase_ , UpperCamelCase_): r += F"{t * (self._level)}{v}\n" self._level += 1 else: r += F"{t * (self._level)}{k}: {v} ({type(UpperCamelCase_).__name__})\n" __UpperCAmelCase : Tuple = level return r[:-1] @classmethod def a_ ( cls : List[Any] , UpperCamelCase_ : str , **UpperCamelCase_ : Dict): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = cls.get_config_dict(UpperCamelCase_ , **UpperCamelCase_) return cls(UpperCamelCase_) @classmethod def a_ ( cls : Dict , UpperCamelCase_ : str , **UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : Tuple = kwargs.pop("cache_dir" , UpperCamelCase_) __UpperCAmelCase : Optional[int] = kwargs.pop("force_download" , UpperCamelCase_) __UpperCAmelCase : Optional[int] = kwargs.pop("resume_download" , UpperCamelCase_) __UpperCAmelCase : int = kwargs.pop("proxies" , UpperCamelCase_) __UpperCAmelCase : Tuple = kwargs.pop("local_files_only" , UpperCamelCase_) if os.path.isdir(UpperCamelCase_): __UpperCAmelCase : Tuple = os.path.join(UpperCamelCase_ , UpperCamelCase_) elif os.path.isfile(UpperCamelCase_) or is_remote_url(UpperCamelCase_): __UpperCAmelCase : Union[str, Any] = pretrained_model_name_or_path else: __UpperCAmelCase : Union[str, Any] = hf_bucket_url(UpperCamelCase_ , filename=UpperCamelCase_ , use_cdn=UpperCamelCase_) try: # Load from URL or cache if already cached __UpperCAmelCase : Tuple = cached_path( UpperCamelCase_ , cache_dir=UpperCamelCase_ , force_download=UpperCamelCase_ , proxies=UpperCamelCase_ , resume_download=UpperCamelCase_ , local_files_only=UpperCamelCase_ , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __UpperCAmelCase : int = Config.load_yaml(UpperCamelCase_) except EnvironmentError: __UpperCAmelCase : Tuple = "Can't load config for" raise EnvironmentError(UpperCamelCase_) if resolved_config_file == config_file: print("loading configuration file from path") else: print("loading configuration file cache") return Config.load_yaml(UpperCamelCase_), kwargs def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" __UpperCAmelCase : int = torch.load("dump.pt" , map_location=in_tensor.device ) __UpperCAmelCase : Dict = in_tensor.numpy() __UpperCAmelCase : List[str] = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(UpperCamelCase , UpperCamelCase , rtol=0.01 , atol=0.1 ), ( f"{sum([1 for x in np.isclose(UpperCamelCase , UpperCamelCase , rtol=0.01 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %" " element-wise mismatch" ) raise Exception("tensors are all good" ) # Hugging face functions below def _UpperCamelCase ( UpperCamelCase ) -> Tuple: """simple docstring""" __UpperCAmelCase : int = urlparse(UpperCamelCase ) return parsed.scheme in ("http", "https") def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase=True ) -> str: """simple docstring""" __UpperCAmelCase : Optional[Any] = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __UpperCAmelCase : List[Any] = "/" not in model_id if legacy_format: return f"{endpoint}/{model_id}-{filename}" else: return f"{endpoint}/{model_id}/{filename}" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase=None , UpperCamelCase=0 , UpperCamelCase=None , ) -> Tuple: """simple docstring""" __UpperCAmelCase : Union[str, Any] = "python/{}".format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(UpperCamelCase , UpperCamelCase ): ua += "; " + "; ".join("{}/{}".format(UpperCamelCase , UpperCamelCase ) for k, v in user_agent.items() ) elif isinstance(UpperCamelCase , UpperCamelCase ): ua += "; " + user_agent __UpperCAmelCase : Dict = {"user-agent": ua} if resume_size > 0: __UpperCAmelCase : str = "bytes=%d-" % (resume_size,) __UpperCAmelCase : Optional[int] = requests.get(UpperCamelCase , stream=UpperCamelCase , proxies=UpperCamelCase , headers=UpperCamelCase ) if response.status_code == 416: # Range not satisfiable return __UpperCAmelCase : str = response.headers.get("Content-Length" ) __UpperCAmelCase : List[Any] = resume_size + int(UpperCamelCase ) if content_length is not None else None __UpperCAmelCase : Any = tqdm( unit="B" , unit_scale=UpperCamelCase , total=UpperCamelCase , initial=UpperCamelCase , desc="Downloading" , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(UpperCamelCase ) ) temp_file.write(UpperCamelCase ) progress.close() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=None , UpperCamelCase=False , UpperCamelCase=None , UpperCamelCase=10 , UpperCamelCase=False , UpperCamelCase=None , UpperCamelCase=False , ) -> Dict: """simple docstring""" if cache_dir is None: __UpperCAmelCase : Dict = TRANSFORMERS_CACHE if isinstance(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : Any = str(UpperCamelCase ) os.makedirs(UpperCamelCase , exist_ok=UpperCamelCase ) __UpperCAmelCase : Optional[Any] = None if not local_files_only: try: __UpperCAmelCase : Optional[Any] = requests.head(UpperCamelCase , allow_redirects=UpperCamelCase , proxies=UpperCamelCase , timeout=UpperCamelCase ) if response.status_code == 200: __UpperCAmelCase : Optional[int] = response.headers.get("ETag" ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __UpperCAmelCase : Any = url_to_filename(UpperCamelCase , UpperCamelCase ) # get cache path to put the file __UpperCAmelCase : int = os.path.join(UpperCamelCase , UpperCamelCase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(UpperCamelCase ): return cache_path else: __UpperCAmelCase : Dict = [ file for file in fnmatch.filter(os.listdir(UpperCamelCase ) , filename + ".*" ) if not file.endswith(".json" ) and not file.endswith(".lock" ) ] if len(UpperCamelCase ) > 0: return os.path.join(UpperCamelCase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( "Cannot find the requested files in the cached path and outgoing traffic has been" " disabled. To enable model look-ups and downloads online, set 'local_files_only'" " to False." ) return None # From now on, etag is not None. if os.path.exists(UpperCamelCase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __UpperCAmelCase : Dict = cache_path + ".lock" with FileLock(UpperCamelCase ): # If the download just completed while the lock was activated. if os.path.exists(UpperCamelCase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __UpperCAmelCase : List[Any] = cache_path + ".incomplete" @contextmanager def _resumable_file_manager(): with open(UpperCamelCase , "a+b" ) as f: yield f __UpperCAmelCase : str = _resumable_file_manager if os.path.exists(UpperCamelCase ): __UpperCAmelCase : Dict = os.stat(UpperCamelCase ).st_size else: __UpperCAmelCase : List[str] = 0 else: __UpperCAmelCase : int = partial(tempfile.NamedTemporaryFile , dir=UpperCamelCase , delete=UpperCamelCase ) __UpperCAmelCase : List[Any] = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( "%s not found in cache or force_download set to True, downloading to %s" , UpperCamelCase , temp_file.name , ) http_get( UpperCamelCase , UpperCamelCase , proxies=UpperCamelCase , resume_size=UpperCamelCase , user_agent=UpperCamelCase , ) os.replace(temp_file.name , UpperCamelCase ) __UpperCAmelCase : List[Any] = {"url": url, "etag": etag} __UpperCAmelCase : str = cache_path + ".json" with open(UpperCamelCase , "w" ) as meta_file: json.dump(UpperCamelCase , UpperCamelCase ) return cache_path def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=None ) -> List[str]: """simple docstring""" __UpperCAmelCase : int = url.encode("utf-8" ) __UpperCAmelCase : List[str] = shaaaa(UpperCamelCase ) __UpperCAmelCase : Dict = url_hash.hexdigest() if etag: __UpperCAmelCase : List[Any] = etag.encode("utf-8" ) __UpperCAmelCase : str = shaaaa(UpperCamelCase ) filename += "." + etag_hash.hexdigest() if url.endswith(".h5" ): filename += ".h5" return filename def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=None , UpperCamelCase=False , UpperCamelCase=None , UpperCamelCase=False , UpperCamelCase=None , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=False , ) -> Optional[int]: """simple docstring""" if cache_dir is None: __UpperCAmelCase : Dict = TRANSFORMERS_CACHE if isinstance(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[Any] = str(UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : Any = str(UpperCamelCase ) if is_remote_url(UpperCamelCase ): # URL, so get it from the cache (downloading if necessary) __UpperCAmelCase : List[Any] = get_from_cache( UpperCamelCase , cache_dir=UpperCamelCase , force_download=UpperCamelCase , proxies=UpperCamelCase , resume_download=UpperCamelCase , user_agent=UpperCamelCase , local_files_only=UpperCamelCase , ) elif os.path.exists(UpperCamelCase ): # File, and it exists. __UpperCAmelCase : List[str] = url_or_filename elif urlparse(UpperCamelCase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError("file {} not found".format(UpperCamelCase ) ) else: # Something unknown raise ValueError("unable to parse {} as a URL or as a local path".format(UpperCamelCase ) ) if extract_compressed_file: if not is_zipfile(UpperCamelCase ) and not tarfile.is_tarfile(UpperCamelCase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = os.path.split(UpperCamelCase ) __UpperCAmelCase : Dict = output_file.replace("." , "-" ) + "-extracted" __UpperCAmelCase : int = os.path.join(UpperCamelCase , UpperCamelCase ) if os.path.isdir(UpperCamelCase ) and os.listdir(UpperCamelCase ) and not force_extract: return output_path_extracted # Prevent parallel extractions __UpperCAmelCase : List[Any] = output_path + ".lock" with FileLock(UpperCamelCase ): shutil.rmtree(UpperCamelCase , ignore_errors=UpperCamelCase ) os.makedirs(UpperCamelCase ) if is_zipfile(UpperCamelCase ): with ZipFile(UpperCamelCase , "r" ) as zip_file: zip_file.extractall(UpperCamelCase ) zip_file.close() elif tarfile.is_tarfile(UpperCamelCase ): __UpperCAmelCase : Optional[Any] = tarfile.open(UpperCamelCase ) tar_file.extractall(UpperCamelCase ) tar_file.close() else: raise EnvironmentError("Archive format of {} could not be identified".format(UpperCamelCase ) ) return output_path_extracted return output_path def _UpperCamelCase ( UpperCamelCase , UpperCamelCase="," ) -> Union[str, Any]: """simple docstring""" assert isinstance(UpperCamelCase , UpperCamelCase ) if os.path.isfile(UpperCamelCase ): with open(UpperCamelCase ) as f: __UpperCAmelCase : int = eval(f.read() ) else: __UpperCAmelCase : str = requests.get(UpperCamelCase ) try: __UpperCAmelCase : List[Any] = requests.json() except Exception: __UpperCAmelCase : List[Any] = req.content.decode() assert data is not None, "could not connect" try: __UpperCAmelCase : Dict = eval(UpperCamelCase ) except Exception: __UpperCAmelCase : List[str] = data.split("\n" ) req.close() return data def _UpperCamelCase ( UpperCamelCase ) -> List[str]: """simple docstring""" __UpperCAmelCase : Union[str, Any] = requests.get(UpperCamelCase ) __UpperCAmelCase : List[Any] = np.array(Image.open(BytesIO(response.content ) ) ) return img def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : int = url.split("/" )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(UpperCamelCase ) with open(UpperCamelCase , "rb" ) as stream: __UpperCAmelCase : Dict = pkl.load(UpperCamelCase ) __UpperCAmelCase : Any = weights.pop("model" ) __UpperCAmelCase : List[Any] = {} for k, v in model.items(): __UpperCAmelCase : int = torch.from_numpy(UpperCamelCase ) if "running_var" in k: __UpperCAmelCase : Any = torch.tensor([0] ) __UpperCAmelCase : Any = k.replace("running_var" , "num_batches_tracked" ) __UpperCAmelCase : Optional[int] = zero return new def _UpperCamelCase ( ) -> Optional[int]: """simple docstring""" print(f"{os.path.abspath(os.path.join(UpperCamelCase , os.pardir ) )}/demo.ipynb" ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase="RGB" ) -> Tuple: """simple docstring""" assert isinstance(UpperCamelCase , UpperCamelCase ) if os.path.isfile(UpperCamelCase ): __UpperCAmelCase : List[str] = cva.imread(UpperCamelCase ) else: __UpperCAmelCase : List[str] = get_image_from_url(UpperCamelCase ) assert img is not None, f"could not connect to: {im}" __UpperCAmelCase : Union[str, Any] = cva.cvtColor(UpperCamelCase , cva.COLOR_BGR2RGB ) if input_format == "RGB": __UpperCAmelCase : Tuple = img[:, :, ::-1] return img def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=1 ) -> Dict: """simple docstring""" return (images[i : i + batch] for i in range(0 , len(UpperCamelCase ) , UpperCamelCase ))
77
"""simple docstring""" from scipy.stats import spearmanr import datasets A = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ A = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ A = r"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def a_ ( self : Any): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float"), "references": datasets.Value("float"), }) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def a_ ( self : Union[str, Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : int=False): """simple docstring""" __UpperCAmelCase : List[str] = spearmanr(UpperCamelCase_ , UpperCamelCase_) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
77
1
"""simple docstring""" from scipy.stats import spearmanr import datasets A = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ A = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ A = r"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def a_ ( self : Any): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float"), "references": datasets.Value("float"), }) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def a_ ( self : Union[str, Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : int=False): """simple docstring""" __UpperCAmelCase : List[str] = spearmanr(UpperCamelCase_ , UpperCamelCase_) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
77
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) A = {"""vocab_file""": """spiece.model"""} A = { """vocab_file""": { """bert_for_seq_generation""": ( """https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model""" ), } } A = {"""bert_for_seq_generation""": 512} class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = [] lowercase_ = ["input_ids", "attention_mask"] def __init__( self : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : List[str]="<s>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Optional[int]="<unk>" , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : List[Any]="<::::>" , UpperCamelCase_ : Optional[Dict[str, Any]] = None , **UpperCamelCase_ : List[Any] , ): """simple docstring""" __UpperCAmelCase : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) __UpperCAmelCase : Dict = vocab_file __UpperCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(UpperCamelCase_) @property def a_ ( self : List[str]): """simple docstring""" return self.sp_model.get_piece_size() def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : int = {self.convert_ids_to_tokens(UpperCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self : int): """simple docstring""" __UpperCAmelCase : Optional[int] = self.__dict__.copy() __UpperCAmelCase : List[Any] = None return state def __setstate__( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): __UpperCAmelCase : List[Any] = {} __UpperCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def a_ ( self : Any , UpperCamelCase_ : str): """simple docstring""" return self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_) def a_ ( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" return self.sp_model.piece_to_id(UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : int = self.sp_model.IdToPiece(UpperCamelCase_) return token def a_ ( self : Dict , UpperCamelCase_ : Optional[Any]): """simple docstring""" __UpperCAmelCase : int = [] __UpperCAmelCase : Tuple = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCamelCase_) + token __UpperCAmelCase : List[Any] = [] else: current_sub_tokens.append(UpperCamelCase_) out_string += self.sp_model.decode(UpperCamelCase_) return out_string.strip() def a_ ( self : List[str] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : Tuple = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , UpperCamelCase_) elif not os.path.isfile(self.vocab_file): with open(UpperCamelCase_ , "wb") as fi: __UpperCAmelCase : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json""", # See all REALM models at https://huggingface.co/models?filter=realm } class a__ ( __magic_name__ ): lowercase_ = "realm" def __init__( self : List[str] , UpperCamelCase_ : int=30522 , UpperCamelCase_ : List[str]=768 , UpperCamelCase_ : Optional[Any]=128 , UpperCamelCase_ : Any=12 , UpperCamelCase_ : Optional[int]=12 , UpperCamelCase_ : Dict=8 , UpperCamelCase_ : Any=3072 , UpperCamelCase_ : Any="gelu_new" , UpperCamelCase_ : List[Any]=0.1 , UpperCamelCase_ : Optional[int]=0.1 , UpperCamelCase_ : Optional[Any]=512 , UpperCamelCase_ : int=2 , UpperCamelCase_ : Union[str, Any]=0.02 , UpperCamelCase_ : Optional[int]=1e-12 , UpperCamelCase_ : Optional[int]=256 , UpperCamelCase_ : Optional[int]=10 , UpperCamelCase_ : Any=1e-3 , UpperCamelCase_ : Optional[Any]=5 , UpperCamelCase_ : List[Any]=320 , UpperCamelCase_ : Tuple=13353718 , UpperCamelCase_ : Optional[Any]=5000 , UpperCamelCase_ : Union[str, Any]=1 , UpperCamelCase_ : Optional[int]=0 , UpperCamelCase_ : Union[str, Any]=2 , **UpperCamelCase_ : Dict , ): """simple docstring""" super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_) # Common config __UpperCAmelCase : str = vocab_size __UpperCAmelCase : List[Any] = max_position_embeddings __UpperCAmelCase : int = hidden_size __UpperCAmelCase : Tuple = retriever_proj_size __UpperCAmelCase : Any = num_hidden_layers __UpperCAmelCase : Union[str, Any] = num_attention_heads __UpperCAmelCase : int = num_candidates __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : int = hidden_act __UpperCAmelCase : List[Any] = hidden_dropout_prob __UpperCAmelCase : Optional[Any] = attention_probs_dropout_prob __UpperCAmelCase : Union[str, Any] = initializer_range __UpperCAmelCase : List[str] = type_vocab_size __UpperCAmelCase : Dict = layer_norm_eps # Reader config __UpperCAmelCase : Union[str, Any] = span_hidden_size __UpperCAmelCase : str = max_span_width __UpperCAmelCase : Union[str, Any] = reader_layer_norm_eps __UpperCAmelCase : str = reader_beam_size __UpperCAmelCase : Any = reader_seq_len # Retrieval config __UpperCAmelCase : Dict = num_block_records __UpperCAmelCase : List[str] = searcher_beam_size
77
"""simple docstring""" import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed A = """true""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=16 ) -> Tuple: """simple docstring""" set_seed(42 ) __UpperCAmelCase : Dict = RegressionModel() __UpperCAmelCase : Optional[Any] = deepcopy(UpperCamelCase ) __UpperCAmelCase : Any = RegressionDataset(length=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = DataLoader(UpperCamelCase , batch_size=UpperCamelCase ) model.to(accelerator.device ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return model, ddp_model, dataloader def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=False ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) __UpperCAmelCase : Dict = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(UpperCamelCase ): __UpperCAmelCase : Dict = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=UpperCamelCase , max_length=UpperCamelCase ) return outputs with accelerator.main_process_first(): __UpperCAmelCase : str = dataset.map( UpperCamelCase , batched=UpperCamelCase , remove_columns=["idx", "sentence1", "sentence2"] , ) __UpperCAmelCase : List[Any] = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(UpperCamelCase ): if use_longest: return tokenizer.pad(UpperCamelCase , padding="longest" , return_tensors="pt" ) return tokenizer.pad(UpperCamelCase , padding="max_length" , max_length=128 , return_tensors="pt" ) return DataLoader(UpperCamelCase , shuffle=UpperCamelCase , collate_fn=UpperCamelCase , batch_size=16 ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[Any] = Accelerator(dispatch_batches=UpperCamelCase , split_batches=UpperCamelCase ) __UpperCAmelCase : int = get_dataloader(UpperCamelCase , not dispatch_batches ) __UpperCAmelCase : Any = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Dict = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Dict = [] for batch in dataloader: __UpperCAmelCase , __UpperCAmelCase : int = batch.values() with torch.no_grad(): __UpperCAmelCase : int = model(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : List[str] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = [], [] for logit, targ in logits_and_targets: logits.append(UpperCamelCase ) targs.append(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = torch.cat(UpperCamelCase ), torch.cat(UpperCamelCase ) return logits, targs def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=16 ) -> int: """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = get_basic_setup(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = generate_predictions(UpperCamelCase , UpperCamelCase , UpperCamelCase ) assert ( len(UpperCamelCase ) == num_samples ), f"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(UpperCamelCase )}" def _UpperCamelCase ( UpperCamelCase = False , UpperCamelCase = False ) -> List[str]: """simple docstring""" __UpperCAmelCase : List[str] = evaluate.load("glue" , "mrpc" ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = get_mrpc_setup(UpperCamelCase , UpperCamelCase ) # First do baseline __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = setup["no"] model.to(UpperCamelCase ) model.eval() for batch in dataloader: batch.to(UpperCamelCase ) with torch.inference_mode(): __UpperCAmelCase : List[str] = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=UpperCamelCase , references=batch["labels"] ) __UpperCAmelCase : str = metric.compute() # Then do distributed __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): __UpperCAmelCase : Any = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) __UpperCAmelCase : Union[str, Any] = batch["labels"] __UpperCAmelCase , __UpperCAmelCase : Any = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=UpperCamelCase , references=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def _UpperCamelCase ( ) -> List[Any]: """simple docstring""" __UpperCAmelCase : Dict = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(UpperCamelCase , UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __UpperCAmelCase : Union[str, Any] = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(UpperCamelCase , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) __UpperCAmelCase : Any = Accelerator() test_torch_metrics(UpperCamelCase , 512 ) accelerator.state._reset_state() def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
77
1
"""simple docstring""" import copy from typing import Dict, List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING A = { """facebook/mask2former-swin-small-coco-instance""": ( """https://huggingface.co/facebook/mask2former-swin-small-coco-instance/blob/main/config.json""" ) # See all Mask2Former models at https://huggingface.co/models?filter=mask2former } A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = "mask2former" lowercase_ = ["swin"] lowercase_ = {"hidden_size": "hidden_dim"} def __init__( self : Any , UpperCamelCase_ : Optional[Dict] = None , UpperCamelCase_ : int = 256 , UpperCamelCase_ : int = 256 , UpperCamelCase_ : int = 256 , UpperCamelCase_ : int = 1024 , UpperCamelCase_ : str = "relu" , UpperCamelCase_ : int = 6 , UpperCamelCase_ : int = 10 , UpperCamelCase_ : int = 8 , UpperCamelCase_ : float = 0.0 , UpperCamelCase_ : int = 2048 , UpperCamelCase_ : bool = False , UpperCamelCase_ : bool = False , UpperCamelCase_ : int = 4 , UpperCamelCase_ : int = 255 , UpperCamelCase_ : int = 100 , UpperCamelCase_ : float = 0.1 , UpperCamelCase_ : float = 2.0 , UpperCamelCase_ : float = 5.0 , UpperCamelCase_ : float = 5.0 , UpperCamelCase_ : int = 12544 , UpperCamelCase_ : float = 3.0 , UpperCamelCase_ : float = 0.75 , UpperCamelCase_ : float = 0.02 , UpperCamelCase_ : float = 1.0 , UpperCamelCase_ : bool = True , UpperCamelCase_ : List[int] = [4, 8, 16, 32] , UpperCamelCase_ : bool = None , **UpperCamelCase_ : int , ): """simple docstring""" if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `Swin` backbone.") __UpperCAmelCase : List[Any] = CONFIG_MAPPING["swin"]( image_size=224 , in_channels=3 , patch_size=4 , embed_dim=96 , depths=[2, 2, 18, 2] , num_heads=[3, 6, 12, 24] , window_size=7 , drop_path_rate=0.3 , use_absolute_embeddings=UpperCamelCase_ , out_features=["stage1", "stage2", "stage3", "stage4"] , ) if isinstance(UpperCamelCase_ , UpperCamelCase_): __UpperCAmelCase : Dict = backbone_config.pop("model_type") __UpperCAmelCase : Optional[int] = CONFIG_MAPPING[backbone_model_type] __UpperCAmelCase : Dict = config_class.from_dict(UpperCamelCase_) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"Backbone {backbone_config.model_type} is not a supported model and may not be compatible with Mask2Former. " F"Supported model types: {','.join(self.backbones_supported)}") __UpperCAmelCase : Optional[int] = backbone_config __UpperCAmelCase : Union[str, Any] = feature_size __UpperCAmelCase : Dict = mask_feature_size __UpperCAmelCase : List[Any] = hidden_dim __UpperCAmelCase : Tuple = encoder_feedforward_dim __UpperCAmelCase : Tuple = activation_function __UpperCAmelCase : Optional[Any] = encoder_layers __UpperCAmelCase : Optional[int] = decoder_layers __UpperCAmelCase : Optional[Any] = num_attention_heads __UpperCAmelCase : Union[str, Any] = dropout __UpperCAmelCase : Any = dim_feedforward __UpperCAmelCase : Tuple = pre_norm __UpperCAmelCase : Any = enforce_input_projection __UpperCAmelCase : Union[str, Any] = common_stride __UpperCAmelCase : List[str] = ignore_value __UpperCAmelCase : Optional[Any] = num_queries __UpperCAmelCase : Optional[Any] = no_object_weight __UpperCAmelCase : Tuple = class_weight __UpperCAmelCase : List[Any] = mask_weight __UpperCAmelCase : int = dice_weight __UpperCAmelCase : int = train_num_points __UpperCAmelCase : Optional[Any] = oversample_ratio __UpperCAmelCase : Dict = importance_sample_ratio __UpperCAmelCase : Tuple = init_std __UpperCAmelCase : List[Any] = init_xavier_std __UpperCAmelCase : str = use_auxiliary_loss __UpperCAmelCase : Any = feature_strides __UpperCAmelCase : Optional[int] = output_auxiliary_logits __UpperCAmelCase : str = decoder_layers super().__init__(**UpperCamelCase_) @classmethod def a_ ( cls : List[str] , UpperCamelCase_ : PretrainedConfig , **UpperCamelCase_ : Optional[int]): """simple docstring""" return cls( backbone_config=UpperCamelCase_ , **UpperCamelCase_ , ) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : str = copy.deepcopy(self.__dict__) __UpperCAmelCase : Optional[Any] = self.backbone_config.to_dict() __UpperCAmelCase : Tuple = self.__class__.model_type return output
77
"""simple docstring""" import math def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = 0 , UpperCamelCase = 0 ) -> list: """simple docstring""" __UpperCAmelCase : Union[str, Any] = end or len(UpperCamelCase ) for i in range(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[Any] = i __UpperCAmelCase : Any = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: __UpperCAmelCase : Dict = array[temp_index - 1] temp_index -= 1 __UpperCAmelCase : str = temp_index_value return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: # Max Heap """simple docstring""" __UpperCAmelCase : Optional[Any] = index __UpperCAmelCase : List[str] = 2 * index + 1 # Left Node __UpperCAmelCase : Union[str, Any] = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: __UpperCAmelCase : Tuple = left_index if right_index < heap_size and array[largest] < array[right_index]: __UpperCAmelCase : int = right_index if largest != index: __UpperCAmelCase , __UpperCAmelCase : List[str] = array[largest], array[index] heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" __UpperCAmelCase : List[Any] = len(UpperCamelCase ) for i in range(n // 2 , -1 , -1 ): heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) for i in range(n - 1 , 0 , -1 ): __UpperCAmelCase , __UpperCAmelCase : int = array[0], array[i] heapify(UpperCamelCase , 0 , UpperCamelCase ) return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Optional[Any] = low __UpperCAmelCase : List[str] = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i __UpperCAmelCase , __UpperCAmelCase : Optional[int] = array[j], array[i] i += 1 def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" if len(UpperCamelCase ) == 0: return array __UpperCAmelCase : Optional[int] = 2 * math.ceil(math.loga(len(UpperCamelCase ) ) ) __UpperCAmelCase : List[Any] = 16 return intro_sort(UpperCamelCase , 0 , len(UpperCamelCase ) , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> list: """simple docstring""" while end - start > size_threshold: if max_depth == 0: return heap_sort(UpperCamelCase ) max_depth -= 1 __UpperCAmelCase : List[Any] = median_of_a(UpperCamelCase , UpperCamelCase , start + ((end - start) // 2) + 1 , end - 1 ) __UpperCAmelCase : Union[str, Any] = partition(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) intro_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[Any] = p return insertion_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() A = input("""Enter numbers separated by a comma : """).strip() A = [float(item) for item in user_input.split(""",""")] print(sort(unsorted))
77
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import is_flaky, require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DonutImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Union[str, Any] , UpperCamelCase_ : Dict , UpperCamelCase_ : int=7 , UpperCamelCase_ : Dict=3 , UpperCamelCase_ : List[str]=18 , UpperCamelCase_ : Union[str, Any]=30 , UpperCamelCase_ : int=400 , UpperCamelCase_ : List[str]=True , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Any=True , UpperCamelCase_ : List[str]=False , UpperCamelCase_ : Dict=True , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : List[Any]=[0.5, 0.5, 0.5] , UpperCamelCase_ : List[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __UpperCAmelCase : Tuple = parent __UpperCAmelCase : Dict = batch_size __UpperCAmelCase : List[Any] = num_channels __UpperCAmelCase : Tuple = image_size __UpperCAmelCase : Optional[Any] = min_resolution __UpperCAmelCase : Optional[Any] = max_resolution __UpperCAmelCase : List[Any] = do_resize __UpperCAmelCase : Tuple = size if size is not None else {"height": 18, "width": 20} __UpperCAmelCase : List[Any] = do_thumbnail __UpperCAmelCase : str = do_align_axis __UpperCAmelCase : int = do_pad __UpperCAmelCase : Optional[int] = do_normalize __UpperCAmelCase : List[Any] = image_mean __UpperCAmelCase : int = image_std def a_ ( self : List[str]): """simple docstring""" return { "do_resize": self.do_resize, "size": self.size, "do_thumbnail": self.do_thumbnail, "do_align_long_axis": self.do_align_axis, "do_pad": self.do_pad, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = DonutImageProcessor if is_vision_available() else None def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : str = DonutImageProcessingTester(self) @property def a_ ( self : List[Any]): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) self.assertTrue(hasattr(UpperCamelCase_ , "do_thumbnail")) self.assertTrue(hasattr(UpperCamelCase_ , "do_align_long_axis")) self.assertTrue(hasattr(UpperCamelCase_ , "do_pad")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCamelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCamelCase_ , "image_std")) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"height": 18, "width": 20}) __UpperCAmelCase : str = self.image_processing_class.from_dict(self.image_processor_dict , size=42) self.assertEqual(image_processor.size , {"height": 42, "width": 42}) # Previous config had dimensions in (width, height) order __UpperCAmelCase : Tuple = self.image_processing_class.from_dict(self.image_processor_dict , size=(42, 84)) self.assertEqual(image_processor.size , {"height": 84, "width": 42}) def a_ ( self : Optional[Any]): """simple docstring""" pass @is_flaky() def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : List[Any] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) @is_flaky() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : Dict = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : List[str] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) @is_flaky() def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : str = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : str = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : List[Any] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
77
"""simple docstring""" import numpy as np from PIL import Image def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : str = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Any = 0 __UpperCAmelCase : Dict = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Tuple = 0 # compute the shape of the output matrix __UpperCAmelCase : Optional[int] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __UpperCAmelCase : List[str] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __UpperCAmelCase : str = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : int = 0 __UpperCAmelCase : int = 0 return updated_arr def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : List[str] = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Any = 0 # compute the shape of the output matrix __UpperCAmelCase : Tuple = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __UpperCAmelCase : str = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __UpperCAmelCase : Tuple = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Optional[Any] = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name="""avgpooling""", verbose=True) # Loading the image A = Image.open("""path_to_image""") # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase = 1 , UpperCamelCase = 1000 ) -> int: """simple docstring""" __UpperCAmelCase : int = 1 __UpperCAmelCase : Tuple = 0 for divide_by_number in range(UpperCamelCase , digit + 1 ): __UpperCAmelCase : list[int] = [] __UpperCAmelCase : int = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(UpperCamelCase ): __UpperCAmelCase : Any = len(UpperCamelCase ) __UpperCAmelCase : Optional[int] = divide_by_number else: has_been_divided.append(UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: A = None A = logging.get_logger(__name__) A = """▁""" A = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} A = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } A = { """google/pegasus-xsum""": 512, } class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = PegasusTokenizer lowercase_ = ["input_ids", "attention_mask"] def __init__( self : str , UpperCamelCase_ : str=None , UpperCamelCase_ : Optional[int]=None , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Any="<unk>" , UpperCamelCase_ : Tuple="<mask_2>" , UpperCamelCase_ : Any="<mask_1>" , UpperCamelCase_ : Tuple=None , UpperCamelCase_ : str=103 , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" __UpperCAmelCase : Optional[int] = offset if additional_special_tokens is not None: if not isinstance(UpperCamelCase_ , UpperCamelCase_): raise TypeError( F"additional_special_tokens should be of type {type(UpperCamelCase_)}, but is" F" {type(UpperCamelCase_)}") __UpperCAmelCase : Any = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"<unk_{i}>" for i in range(len(UpperCamelCase_) , self.offset - 1) ] if len(set(UpperCamelCase_)) != len(UpperCamelCase_): raise ValueError( "Please make sure that the provided additional_special_tokens do not contain an incorrectly" F" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.") __UpperCAmelCase : str = additional_special_tokens_extended else: __UpperCAmelCase : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"<unk_{i}>" for i in range(2 , self.offset)] super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , pad_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , mask_token_sent=UpperCamelCase_ , offset=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Optional[int] = vocab_file __UpperCAmelCase : List[str] = False if not self.vocab_file else True def a_ ( self : Union[str, Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : int = set(self.all_special_ids) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens) + 3)): raise ValueError( "There should be 3 special tokens: mask_token, pad_token, and eos_token +" F" {len(self.additional_special_tokens)} additional_special_tokens, but got {all_special_ids}") return [1 if x in all_special_ids else 0 for x in seq] def a_ ( self : Union[str, Any] , UpperCamelCase_ : List , UpperCamelCase_ : Optional[List] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return self._special_token_mask(UpperCamelCase_) elif token_ids_a is None: return self._special_token_mask(UpperCamelCase_) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a) + [1] def a_ ( self : int , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : List[Any]=None): """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a_ ( self : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : List[str] = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_): copyfile(self.vocab_file , UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> Optional[int]: """simple docstring""" return [ { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], }, { 0: [6], 1: [9], 2: [4, 5], 3: [4], 4: [2, 3], 5: [2], 6: [0, 7], 7: [6], 8: [], 9: [1], }, { 0: [4], 1: [6], 2: [], 3: [5, 6, 7], 4: [0, 6], 5: [3, 8, 9], 6: [1, 3, 4, 7], 7: [3, 6, 8, 9], 8: [5, 7], 9: [5, 7], }, { 0: [1, 3], 1: [0, 2, 4], 2: [1, 3, 4], 3: [0, 2, 4], 4: [1, 2, 3], }, ][index] def _UpperCamelCase ( UpperCamelCase ) -> list[tuple[int, int]]: """simple docstring""" __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : Any = len(UpperCamelCase ) # No of vertices in graph __UpperCAmelCase : Union[str, Any] = [0] * n __UpperCAmelCase : List[Any] = [False] * n def dfs(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[str] = True __UpperCAmelCase : Union[str, Any] = id_ id_ += 1 for to in graph[at]: if to == parent: pass elif not visited[to]: dfs(UpperCamelCase , UpperCamelCase , UpperCamelCase , id_ ) __UpperCAmelCase : Tuple = min(low[at] , low[to] ) if id_ <= low[to]: bridges.append((at, to) if at < to else (to, at) ) else: # This edge is a back edge and cannot be a bridge __UpperCAmelCase : List[Any] = min(low[at] , low[to] ) __UpperCAmelCase : list[tuple[int, int]] = [] for i in range(UpperCamelCase ): if not visited[i]: dfs(UpperCamelCase , -1 , UpperCamelCase , id_ ) return bridges if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __UpperCAmelCase : Optional[Any] = TapasConfig.from_json_file(UpperCamelCase ) # set absolute/relative position embeddings parameter __UpperCAmelCase : Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __UpperCAmelCase : List[str] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WTQ": # run_task_main.py hparams __UpperCAmelCase : Tuple = 4 __UpperCAmelCase : Any = True # hparam_utils.py hparams __UpperCAmelCase : Union[str, Any] = 0.664694 __UpperCAmelCase : Union[str, Any] = 0.207951 __UpperCAmelCase : int = 0.121194 __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : List[str] = True __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : List[str] = 0.0352513 __UpperCAmelCase : Optional[int] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __UpperCAmelCase : int = 4 __UpperCAmelCase : Optional[int] = False # hparam_utils.py hparams __UpperCAmelCase : int = 36.4519 __UpperCAmelCase : str = 0.903421 __UpperCAmelCase : Dict = 222.088 __UpperCAmelCase : Dict = True __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Tuple = True __UpperCAmelCase : Any = 0.763141 __UpperCAmelCase : Optional[Any] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "TABFACT": __UpperCAmelCase : Union[str, Any] = TapasForSequenceClassification(config=UpperCamelCase ) elif task == "MLM": __UpperCAmelCase : Tuple = TapasForMaskedLM(config=UpperCamelCase ) elif task == "INTERMEDIATE_PRETRAINING": __UpperCAmelCase : List[str] = TapasModel(config=UpperCamelCase ) else: raise ValueError(f"Task {task} not supported." ) print(f"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # Save pytorch-model (weights and configuration) print(f"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(UpperCamelCase ) # Save tokenizer files print(f"Save tokenizer files to {pytorch_dump_path}" ) __UpperCAmelCase : str = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + "vocab.txt" , model_max_length=512 ) tokenizer.save_pretrained(UpperCamelCase ) print("Used relative position embeddings:" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
77
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Tuple , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int]=7 , UpperCamelCase_ : Optional[Any]=3 , UpperCamelCase_ : Tuple=18 , UpperCamelCase_ : Tuple=30 , UpperCamelCase_ : List[str]=400 , UpperCamelCase_ : Tuple=True , UpperCamelCase_ : Tuple=None , UpperCamelCase_ : Dict=True , ): """simple docstring""" __UpperCAmelCase : Dict = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : str = batch_size __UpperCAmelCase : Optional[int] = num_channels __UpperCAmelCase : List[str] = image_size __UpperCAmelCase : Tuple = min_resolution __UpperCAmelCase : List[Any] = max_resolution __UpperCAmelCase : int = do_resize __UpperCAmelCase : Any = size __UpperCAmelCase : Tuple = apply_ocr def a_ ( self : Tuple): """simple docstring""" return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = LayoutLMvaImageProcessor if is_pytesseract_available() else None def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Dict = LayoutLMvaImageProcessingTester(self) @property def a_ ( self : Optional[int]): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) self.assertTrue(hasattr(UpperCamelCase_ , "apply_ocr")) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"height": 18, "width": 18}) __UpperCAmelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42) self.assertEqual(image_processor.size , {"height": 42, "width": 42}) def a_ ( self : Optional[Any]): """simple docstring""" pass def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Any = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Optional[int] = image_processing(image_inputs[0] , return_tensors="pt") self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) self.assertIsInstance(encoding.words , UpperCamelCase_) self.assertIsInstance(encoding.boxes , UpperCamelCase_) # Test batched __UpperCAmelCase : Optional[Any] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : List[Any] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : int = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : List[str] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : Tuple = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Union[str, Any] = image_processing(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Any = LayoutLMvaImageProcessor() from datasets import load_dataset __UpperCAmelCase : List[str] = load_dataset("hf-internal-testing/fixtures_docvqa" , split="test") __UpperCAmelCase : Optional[Any] = Image.open(ds[0]["file"]).convert("RGB") __UpperCAmelCase : List[str] = image_processing(UpperCamelCase_ , return_tensors="pt") self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224)) self.assertEqual(len(encoding.words) , len(encoding.boxes)) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 __UpperCAmelCase : Optional[int] = [["11:14", "to", "11:39", "a.m", "11:39", "to", "11:44", "a.m.", "11:44", "a.m.", "to", "12:25", "p.m.", "12:25", "to", "12:58", "p.m.", "12:58", "to", "4:00", "p.m.", "2:00", "to", "5:00", "p.m.", "Coffee", "Break", "Coffee", "will", "be", "served", "for", "men", "and", "women", "in", "the", "lobby", "adjacent", "to", "exhibit", "area.", "Please", "move", "into", "exhibit", "area.", "(Exhibits", "Open)", "TRRF", "GENERAL", "SESSION", "(PART", "|)", "Presiding:", "Lee", "A.", "Waller", "TRRF", "Vice", "President", "“Introductory", "Remarks”", "Lee", "A.", "Waller,", "TRRF", "Vice", "Presi-", "dent", "Individual", "Interviews", "with", "TRRF", "Public", "Board", "Members", "and", "Sci-", "entific", "Advisory", "Council", "Mem-", "bers", "Conducted", "by", "TRRF", "Treasurer", "Philip", "G.", "Kuehn", "to", "get", "answers", "which", "the", "public", "refrigerated", "warehousing", "industry", "is", "looking", "for.", "Plus", "questions", "from", "the", "floor.", "Dr.", "Emil", "M.", "Mrak,", "University", "of", "Cal-", "ifornia,", "Chairman,", "TRRF", "Board;", "Sam", "R.", "Cecil,", "University", "of", "Georgia", "College", "of", "Agriculture;", "Dr.", "Stanley", "Charm,", "Tufts", "University", "School", "of", "Medicine;", "Dr.", "Robert", "H.", "Cotton,", "ITT", "Continental", "Baking", "Company;", "Dr.", "Owen", "Fennema,", "University", "of", "Wis-", "consin;", "Dr.", "Robert", "E.", "Hardenburg,", "USDA.", "Questions", "and", "Answers", "Exhibits", "Open", "Capt.", "Jack", "Stoney", "Room", "TRRF", "Scientific", "Advisory", "Council", "Meeting", "Ballroom", "Foyer"]] # noqa: E231 __UpperCAmelCase : List[str] = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , UpperCamelCase_) self.assertListEqual(encoding.boxes , UpperCamelCase_) # with apply_OCR = False __UpperCAmelCase : Dict = LayoutLMvaImageProcessor(apply_ocr=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = image_processing(UpperCamelCase_ , return_tensors="pt") self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224))
77
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = "cpu" , UpperCamelCase = None ) -> None: """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.load(UpperCamelCase , map_location=UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(UpperCamelCase , torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) __UpperCAmelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __UpperCAmelCase : str = src_path torch.save(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A = { """configuration_maskformer""": ["""MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MaskFormerConfig"""], """configuration_maskformer_swin""": ["""MaskFormerSwinConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""MaskFormerFeatureExtractor"""] A = ["""MaskFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """MaskFormerForInstanceSegmentation""", """MaskFormerModel""", """MaskFormerPreTrainedModel""", ] A = [ """MaskFormerSwinBackbone""", """MaskFormerSwinModel""", """MaskFormerSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_maskformer import MASKFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskFormerConfig from .configuration_maskformer_swin import MaskFormerSwinConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_maskformer import MaskFormerFeatureExtractor from .image_processing_maskformer import MaskFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskformer import ( MASKFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskFormerForInstanceSegmentation, MaskFormerModel, MaskFormerPreTrainedModel, ) from .modeling_maskformer_swin import ( MaskFormerSwinBackbone, MaskFormerSwinModel, MaskFormerSwinPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
77
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": A = pd.read_csv("""sample_data.csv""", header=None) A = df.shape[:1][0] # If you're using some other dataset input the target column A = df.iloc[:, 1:2] A = actual_data.values.reshape(len_data, 1) A = MinMaxScaler().fit_transform(actual_data) A = 10 A = 5 A = 20 A = len_data - periods * look_back A = actual_data[:division] A = actual_data[division - look_back :] A , A = [], [] A , A = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) A = np.array(train_x) A = np.array(test_x) A = np.array([list(i.ravel()) for i in train_y]) A = np.array([list(i.ravel()) for i in test_y]) A = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") A = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) A = model.predict(x_test)
77
1
"""simple docstring""" import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def _UpperCamelCase ( ) -> List[str]: """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(UpperCamelCase ): requests.request("GET" , "https://huggingface.co" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("GET" , "https://huggingface.co" , timeout=1.0 ) @pytest.mark.integration def _UpperCamelCase ( ) -> Any: """simple docstring""" with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("GET" , "https://huggingface.co" ) def _UpperCamelCase ( ) -> List[Any]: """simple docstring""" with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(UpperCamelCase ): http_head("https://huggingface.co" )
77
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A = 250_004 A = 250_020 @require_sentencepiece @require_tokenizers class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = MBartTokenizer lowercase_ = MBartTokenizerFast lowercase_ = True lowercase_ = True def a_ ( self : str): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase : Any = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) tokenizer.save_pretrained(self.tmpdirname) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Dict = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer.tokenize("This is a test") self.assertListEqual(UpperCamelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase_) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __UpperCAmelCase : List[Any] = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __UpperCAmelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def a_ ( self : Dict): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __UpperCAmelCase : Dict = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})"): __UpperCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = tempfile.mkdtemp() __UpperCAmelCase : Optional[int] = tokenizer_r.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Any = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) __UpperCAmelCase : Any = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f) self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Tuple = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=True __UpperCAmelCase : Optional[int] = tempfile.mkdtemp() __UpperCAmelCase : Dict = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : int = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=False __UpperCAmelCase : Tuple = tempfile.mkdtemp() __UpperCAmelCase : int = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way __UpperCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : str = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) @require_torch @require_sentencepiece @require_tokenizers class a__ ( unittest.TestCase ): lowercase_ = "facebook/mbart-large-en-ro" lowercase_ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowercase_ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowercase_ = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def a_ ( cls : int): """simple docstring""" __UpperCAmelCase : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en_XX" , tgt_lang="ro_RO") __UpperCAmelCase : Union[str, Any] = 1 return cls def a_ ( self : List[Any]): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ar_AR"] , 250001) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["en_EN"] , 250004) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ro_RO"] , 250020) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" self.assertIn(UpperCamelCase_ , self.tokenizer.all_special_ids) __UpperCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] __UpperCAmelCase : Optional[Any] = self.tokenizer.decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase_) self.assertEqual(UpperCamelCase_ , UpperCamelCase_) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , UpperCamelCase_) __UpperCAmelCase : Tuple = 10 __UpperCAmelCase : List[Any] = self.tokenizer(UpperCamelCase_ , max_length=UpperCamelCase_ , truncation=UpperCamelCase_).input_ids[0] self.assertEqual(ids[-2] , 2) self.assertEqual(ids[-1] , UpperCamelCase_) self.assertEqual(len(UpperCamelCase_) , UpperCamelCase_) def a_ ( self : Any): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"]) , [250026, 250001]) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[str] = tempfile.mkdtemp() __UpperCAmelCase : Union[str, Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase_) __UpperCAmelCase : List[Any] = MBartTokenizer.from_pretrained(UpperCamelCase_) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCamelCase_) @require_torch def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , return_tensors="pt") __UpperCAmelCase : Dict = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=len(self.expected_src_tokens) , return_tensors="pt" , ) __UpperCAmelCase : Tuple = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_) self.assertEqual((2, 14) , batch.input_ids.shape) self.assertEqual((2, 14) , batch.attention_mask.shape) __UpperCAmelCase : List[str] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) self.assertEqual(2 , batch.decoder_input_ids[0, -1]) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , []) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE]) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : List[str] = self.tokenizer(self.src_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=3 , return_tensors="pt") __UpperCAmelCase : Any = self.tokenizer( text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=10 , return_tensors="pt") __UpperCAmelCase : int = targets["input_ids"] __UpperCAmelCase : Any = shift_tokens_right(UpperCamelCase_ , self.tokenizer.pad_token_id) self.assertEqual(batch.input_ids.shape[1] , 3) self.assertEqual(batch.decoder_input_ids.shape[1] , 10) @require_torch def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="ar_AR") self.assertEqual( nested_simplify(UpperCamelCase_) , { # A, test, EOS, en_XX "input_ids": [[62, 3034, 2, 250004]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 250001, } , )
77
1
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation A = logging.get_logger(__name__) A = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} A = { """tokenizer_file""": { """EleutherAI/gpt-neox-20b""": """https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json""", }, } A = { """gpt-neox-20b""": 2_048, } class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , UpperCamelCase_ : Any=None , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : str=None , UpperCamelCase_ : int="<|endoftext|>" , UpperCamelCase_ : Union[str, Any]="<|endoftext|>" , UpperCamelCase_ : Tuple="<|endoftext|>" , UpperCamelCase_ : Optional[int]=False , **UpperCamelCase_ : Optional[int] , ): """simple docstring""" super().__init__( UpperCamelCase_ , UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , unk_token=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , add_prefix_space=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("add_prefix_space" , UpperCamelCase_) != add_prefix_space: __UpperCAmelCase : Optional[Any] = getattr(UpperCamelCase_ , pre_tok_state.pop("type")) __UpperCAmelCase : int = add_prefix_space __UpperCAmelCase : str = pre_tok_class(**UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = add_prefix_space def a_ ( self : List[Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" __UpperCAmelCase : Optional[Any] = self._tokenizer.model.save(UpperCamelCase_ , name=UpperCamelCase_) return tuple(UpperCamelCase_) def a_ ( self : Optional[Any] , UpperCamelCase_ : "Conversation"): """simple docstring""" __UpperCAmelCase : Tuple = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(UpperCamelCase_ , add_special_tokens=UpperCamelCase_) + [self.eos_token_id]) if len(UpperCamelCase_) > self.model_max_length: __UpperCAmelCase : int = input_ids[-self.model_max_length :] return input_ids
77
"""simple docstring""" from typing import Any class a__ : def __init__( self : List[str] , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : str = data __UpperCAmelCase : Optional[Any] = None class a__ : def __init__( self : Any): """simple docstring""" __UpperCAmelCase : Optional[int] = None def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.head while temp is not None: print(temp.data , end=" ") __UpperCAmelCase : Tuple = temp.next print() def a_ ( self : int , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : List[str] = Node(UpperCamelCase_) __UpperCAmelCase : str = self.head __UpperCAmelCase : Optional[int] = new_node def a_ ( self : str , UpperCamelCase_ : List[Any] , UpperCamelCase_ : str): """simple docstring""" if node_data_a == node_data_a: return else: __UpperCAmelCase : int = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Tuple = node_a.next __UpperCAmelCase : List[Any] = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Optional[Any] = node_a.next if node_a is None or node_a is None: return __UpperCAmelCase , __UpperCAmelCase : Any = node_a.data, node_a.data if __name__ == "__main__": A = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
77
1
"""simple docstring""" import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort A = logging.get_logger(__name__) A = { """tensor(bool)""": np.bool_, """tensor(int8)""": np.inta, """tensor(uint8)""": np.uinta, """tensor(int16)""": np.intaa, """tensor(uint16)""": np.uintaa, """tensor(int32)""": np.intaa, """tensor(uint32)""": np.uintaa, """tensor(int64)""": np.intaa, """tensor(uint64)""": np.uintaa, """tensor(float16)""": np.floataa, """tensor(float)""": np.floataa, """tensor(double)""": np.floataa, } class a__ : def __init__( self : Dict , UpperCamelCase_ : Union[str, Any]=None , **UpperCamelCase_ : Optional[Any]): """simple docstring""" logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future.") __UpperCAmelCase : List[Any] = model __UpperCAmelCase : Optional[Any] = kwargs.get("model_save_dir" , UpperCamelCase_) __UpperCAmelCase : Optional[int] = kwargs.get("latest_model_name" , UpperCamelCase_) def __call__( self : Optional[int] , **UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : List[str] = {k: np.array(UpperCamelCase_) for k, v in kwargs.items()} return self.model.run(UpperCamelCase_ , UpperCamelCase_) @staticmethod def a_ ( UpperCamelCase_ : Union[str, Path] , UpperCamelCase_ : str=None , UpperCamelCase_ : str=None): """simple docstring""" if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider") __UpperCAmelCase : Any = "CPUExecutionProvider" return ort.InferenceSession(UpperCamelCase_ , providers=[provider] , sess_options=UpperCamelCase_) def a_ ( self : Dict , UpperCamelCase_ : Union[str, Path] , UpperCamelCase_ : Optional[str] = None , **UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : List[Any] = file_name if file_name is not None else ONNX_WEIGHTS_NAME __UpperCAmelCase : Dict = self.model_save_dir.joinpath(self.latest_model_name) __UpperCAmelCase : List[str] = Path(UpperCamelCase_).joinpath(UpperCamelCase_) try: shutil.copyfile(UpperCamelCase_ , UpperCamelCase_) except shutil.SameFileError: pass # copy external weights (for models >2GB) __UpperCAmelCase : Tuple = self.model_save_dir.joinpath(UpperCamelCase_) if src_path.exists(): __UpperCAmelCase : int = Path(UpperCamelCase_).joinpath(UpperCamelCase_) try: shutil.copyfile(UpperCamelCase_ , UpperCamelCase_) except shutil.SameFileError: pass def a_ ( self : Tuple , UpperCamelCase_ : Union[str, os.PathLike] , **UpperCamelCase_ : List[str] , ): """simple docstring""" if os.path.isfile(UpperCamelCase_): logger.error(F"Provided path ({save_directory}) should be a directory, not a file") return os.makedirs(UpperCamelCase_ , exist_ok=UpperCamelCase_) # saving model weights/files self._save_pretrained(UpperCamelCase_ , **UpperCamelCase_) @classmethod def a_ ( cls : Tuple , UpperCamelCase_ : Union[str, Path] , UpperCamelCase_ : Optional[Union[bool, str, None]] = None , UpperCamelCase_ : Optional[Union[str, None]] = None , UpperCamelCase_ : bool = False , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional["ort.SessionOptions"] = None , **UpperCamelCase_ : List[Any] , ): """simple docstring""" __UpperCAmelCase : List[str] = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(UpperCamelCase_): __UpperCAmelCase : int = OnnxRuntimeModel.load_model( os.path.join(UpperCamelCase_ , UpperCamelCase_) , provider=UpperCamelCase_ , sess_options=UpperCamelCase_) __UpperCAmelCase : Dict = Path(UpperCamelCase_) # load model from hub else: # download model __UpperCAmelCase : Any = hf_hub_download( repo_id=UpperCamelCase_ , filename=UpperCamelCase_ , use_auth_token=UpperCamelCase_ , revision=UpperCamelCase_ , cache_dir=UpperCamelCase_ , force_download=UpperCamelCase_ , ) __UpperCAmelCase : Optional[int] = Path(UpperCamelCase_).parent __UpperCAmelCase : List[str] = Path(UpperCamelCase_).name __UpperCAmelCase : Union[str, Any] = OnnxRuntimeModel.load_model(UpperCamelCase_ , provider=UpperCamelCase_ , sess_options=UpperCamelCase_) return cls(model=UpperCamelCase_ , **UpperCamelCase_) @classmethod def a_ ( cls : List[Any] , UpperCamelCase_ : Union[str, Path] , UpperCamelCase_ : bool = True , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[str] = None , **UpperCamelCase_ : Optional[int] , ): """simple docstring""" __UpperCAmelCase : Optional[int] = None if len(str(UpperCamelCase_).split("@")) == 2: __UpperCAmelCase , __UpperCAmelCase : Any = model_id.split("@") return cls._from_pretrained( model_id=UpperCamelCase_ , revision=UpperCamelCase_ , cache_dir=UpperCamelCase_ , force_download=UpperCamelCase_ , use_auth_token=UpperCamelCase_ , **UpperCamelCase_ , )
77
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore A = """ Human: <<task>> Assistant: """ A = """huggingface-tools/default-prompts""" A = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase="run" ) -> List[str]: """simple docstring""" if prompt_or_repo_id is None: __UpperCAmelCase : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , UpperCamelCase ) is not None: return prompt_or_repo_id __UpperCAmelCase : str = cached_file( UpperCamelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(UpperCamelCase , "r" , encoding="utf-8" ) as f: return f.read()
77
1
"""simple docstring""" from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor A = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def _UpperCamelCase ( UpperCamelCase ) -> Optional[int]: """simple docstring""" if isinstance(UpperCamelCase , torch.Tensor ): return image elif isinstance(UpperCamelCase , PIL.Image.Image ): __UpperCAmelCase : int = [image] __UpperCAmelCase : Optional[Any] = [trans(img.convert("RGB" ) ) for img in image] __UpperCAmelCase : Dict = torch.stack(UpperCamelCase ) return image class a__ ( __magic_name__ ): def __init__( self : Optional[int] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Any): """simple docstring""" super().__init__() # make sure scheduler can always be converted to DDIM __UpperCAmelCase : Any = DDIMScheduler.from_config(scheduler.config) self.register_modules(unet=UpperCamelCase_ , scheduler=UpperCamelCase_) def a_ ( self : Union[str, Any] , UpperCamelCase_ : List[Any]): """simple docstring""" if strength < 0 or strength > 1: raise ValueError(F"The value of strength should in [0.0, 1.0] but is {strength}") def a_ ( self : List[str] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Dict , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Any = min(int(num_inference_steps * strength) , UpperCamelCase_) __UpperCAmelCase : Optional[Any] = max(num_inference_steps - init_timestep , 0) __UpperCAmelCase : Optional[int] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def a_ ( self : Dict , UpperCamelCase_ : Any , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : str=None): """simple docstring""" if not isinstance(UpperCamelCase_ , (torch.Tensor, PIL.Image.Image, list)): raise ValueError( F"`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(UpperCamelCase_)}") __UpperCAmelCase : List[Any] = image.to(device=UpperCamelCase_ , dtype=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) and len(UpperCamelCase_) != batch_size: raise ValueError( F"You have passed a list of generators of length {len(UpperCamelCase_)}, but requested an effective batch" F" size of {batch_size}. Make sure the batch size matches the length of the generators.") __UpperCAmelCase : int = init_latents.shape __UpperCAmelCase : Union[str, Any] = randn_tensor(UpperCamelCase_ , generator=UpperCamelCase_ , device=UpperCamelCase_ , dtype=UpperCamelCase_) # get latents print("add noise to latents at timestep" , UpperCamelCase_) __UpperCAmelCase : str = self.scheduler.add_noise(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : int = init_latents return latents @torch.no_grad() def __call__( self : Any , UpperCamelCase_ : Union[torch.FloatTensor, PIL.Image.Image] = None , UpperCamelCase_ : float = 0.8 , UpperCamelCase_ : int = 1 , UpperCamelCase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCamelCase_ : float = 0.0 , UpperCamelCase_ : int = 50 , UpperCamelCase_ : Optional[bool] = None , UpperCamelCase_ : Optional[str] = "pil" , UpperCamelCase_ : bool = True , ): """simple docstring""" self.check_inputs(UpperCamelCase_) # 2. Preprocess image __UpperCAmelCase : Dict = preprocess(UpperCamelCase_) # 3. set timesteps self.scheduler.set_timesteps(UpperCamelCase_ , device=self.device) __UpperCAmelCase , __UpperCAmelCase : Optional[int] = self.get_timesteps(UpperCamelCase_ , UpperCamelCase_ , self.device) __UpperCAmelCase : List[Any] = timesteps[:1].repeat(UpperCamelCase_) # 4. Prepare latent variables __UpperCAmelCase : Union[str, Any] = self.prepare_latents(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , self.unet.dtype , self.device , UpperCamelCase_) __UpperCAmelCase : List[str] = latents # 5. Denoising loop for t in self.progress_bar(UpperCamelCase_): # 1. predict noise model_output __UpperCAmelCase : Dict = self.unet(UpperCamelCase_ , UpperCamelCase_).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 __UpperCAmelCase : Optional[int] = self.scheduler.step( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , eta=UpperCamelCase_ , use_clipped_model_output=UpperCamelCase_ , generator=UpperCamelCase_ , ).prev_sample __UpperCAmelCase : List[str] = (image / 2 + 0.5).clamp(0 , 1) __UpperCAmelCase : List[str] = image.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": __UpperCAmelCase : Optional[int] = self.numpy_to_pil(UpperCamelCase_) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=UpperCamelCase_)
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available A = { """configuration_ernie""": ["""ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ErnieConfig""", """ErnieOnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST""", """ErnieForCausalLM""", """ErnieForMaskedLM""", """ErnieForMultipleChoice""", """ErnieForNextSentencePrediction""", """ErnieForPreTraining""", """ErnieForQuestionAnswering""", """ErnieForSequenceClassification""", """ErnieForTokenClassification""", """ErnieModel""", """ErniePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar A = TypeVar("""T""") class a__ ( Generic[T] ): lowercase_ = 42 # Cache store of keys lowercase_ = 42 # References of the keys in cache lowercase_ = 1_0 # Maximum capacity of cache def __init__( self : int , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : List[Any] = deque() __UpperCAmelCase : Dict = set() if not n: __UpperCAmelCase : str = sys.maxsize elif n < 0: raise ValueError("n should be an integer greater than 0.") else: __UpperCAmelCase : Dict = n def a_ ( self : List[str] , UpperCamelCase_ : T): """simple docstring""" if x not in self.key_reference: if len(self.dq_store) == LRUCache._MAX_CAPACITY: __UpperCAmelCase : Dict = self.dq_store.pop() self.key_reference.remove(UpperCamelCase_) else: self.dq_store.remove(UpperCamelCase_) self.dq_store.appendleft(UpperCamelCase_) self.key_reference.add(UpperCamelCase_) def a_ ( self : List[str]): """simple docstring""" for k in self.dq_store: print(UpperCamelCase_) def __repr__( self : Tuple): """simple docstring""" return F"LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store)}" if __name__ == "__main__": import doctest doctest.testmod() A = LRUCache(4) lru_cache.refer("""A""") lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer("""A""") lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
77
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class a__ ( nn.Module ): def __init__( self : Union[str, Any]): """simple docstring""" super().__init__() __UpperCAmelCase : Optional[int] = nn.Linear(3 , 4) __UpperCAmelCase : str = nn.BatchNormad(4) __UpperCAmelCase : int = nn.Linear(4 , 5) def a_ ( self : str , UpperCamelCase_ : List[str]): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCamelCase_))) class a__ ( unittest.TestCase ): def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Optional[Any] = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , model.state_dict()) __UpperCAmelCase : Union[str, Any] = os.path.join(UpperCamelCase_ , "index.json") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: __UpperCAmelCase : Optional[int] = os.path.join(UpperCamelCase_ , F"{key}.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on the fact weights are properly loaded def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : int = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: __UpperCAmelCase : List[Any] = torch.randn(2 , 3 , dtype=UpperCamelCase_) with TemporaryDirectory() as tmp_dir: __UpperCAmelCase : Tuple = offload_weight(UpperCamelCase_ , "weight" , UpperCamelCase_ , {}) __UpperCAmelCase : Dict = os.path.join(UpperCamelCase_ , "weight.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) self.assertDictEqual(UpperCamelCase_ , {"weight": {"shape": [2, 3], "dtype": str(UpperCamelCase_).split(".")[1]}}) __UpperCAmelCase : Optional[Any] = load_offloaded_weight(UpperCamelCase_ , index["weight"]) self.assertTrue(torch.equal(UpperCamelCase_ , UpperCamelCase_)) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : List[Any] = ModelForTest() __UpperCAmelCase : Optional[int] = model.state_dict() __UpperCAmelCase : List[str] = {k: v for k, v in state_dict.items() if "linear2" not in k} __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "linear2" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : List[str] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "weight" in k} __UpperCAmelCase : Optional[Any] = {k: v for k, v in state_dict.items() if "weight" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[Any] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) # Duplicates are removed __UpperCAmelCase : str = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Any = {"a.1": 0, "a.10": 1, "a.2": 2} __UpperCAmelCase : Union[str, Any] = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1": 0, "a.2": 2}) __UpperCAmelCase : int = {"a.1.a": 0, "a.10.a": 1, "a.2.a": 2} __UpperCAmelCase : int = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1.a": 0, "a.2.a": 2})
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { """shi-labs/dinat-mini-in1k-224""": """https://huggingface.co/shi-labs/dinat-mini-in1k-224/resolve/main/config.json""", # See all Dinat models at https://huggingface.co/models?filter=dinat } class a__ ( __magic_name__ , __magic_name__ ): lowercase_ = "dinat" lowercase_ = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self : Optional[int] , UpperCamelCase_ : int=4 , UpperCamelCase_ : str=3 , UpperCamelCase_ : Optional[Any]=64 , UpperCamelCase_ : Union[str, Any]=[3, 4, 6, 5] , UpperCamelCase_ : Union[str, Any]=[2, 4, 8, 16] , UpperCamelCase_ : List[Any]=7 , UpperCamelCase_ : List[str]=[[1, 8, 1], [1, 4, 1, 4], [1, 2, 1, 2, 1, 2], [1, 1, 1, 1, 1]] , UpperCamelCase_ : Tuple=3.0 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Dict=0.0 , UpperCamelCase_ : Union[str, Any]=0.0 , UpperCamelCase_ : int=0.1 , UpperCamelCase_ : int="gelu" , UpperCamelCase_ : Dict=0.02 , UpperCamelCase_ : List[Any]=1e-5 , UpperCamelCase_ : List[str]=0.0 , UpperCamelCase_ : int=None , UpperCamelCase_ : Optional[Any]=None , **UpperCamelCase_ : List[Any] , ): """simple docstring""" super().__init__(**UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = patch_size __UpperCAmelCase : int = num_channels __UpperCAmelCase : List[str] = embed_dim __UpperCAmelCase : List[Any] = depths __UpperCAmelCase : List[str] = len(UpperCamelCase_) __UpperCAmelCase : Tuple = num_heads __UpperCAmelCase : Union[str, Any] = kernel_size __UpperCAmelCase : Dict = dilations __UpperCAmelCase : Optional[int] = mlp_ratio __UpperCAmelCase : Tuple = qkv_bias __UpperCAmelCase : Dict = hidden_dropout_prob __UpperCAmelCase : Tuple = attention_probs_dropout_prob __UpperCAmelCase : str = drop_path_rate __UpperCAmelCase : int = hidden_act __UpperCAmelCase : Dict = layer_norm_eps __UpperCAmelCase : List[Any] = initializer_range # we set the hidden_size attribute in order to make Dinat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __UpperCAmelCase : Dict = int(embed_dim * 2 ** (len(UpperCamelCase_) - 1)) __UpperCAmelCase : Dict = layer_scale_init_value __UpperCAmelCase : int = ["stem"] + [F"stage{idx}" for idx in range(1 , len(UpperCamelCase_) + 1)] __UpperCAmelCase , __UpperCAmelCase : str = get_aligned_output_features_output_indices( out_features=UpperCamelCase_ , out_indices=UpperCamelCase_ , stage_names=self.stage_names)
77
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Dict = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __UpperCAmelCase : Union[str, Any] = n - k # Calculate C(n,k) for i in range(UpperCamelCase ): result *= n - i result //= i + 1 return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , UpperCamelCase ) // (node_count + 1) def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if n < 0: raise ValueError("factorial() not defined for negative values" ) __UpperCAmelCase : Optional[Any] = 1 for i in range(1 , n + 1 ): result *= i return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return catalan_number(UpperCamelCase ) * factorial(UpperCamelCase ) if __name__ == "__main__": A = int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
77
1
"""simple docstring""" import os from collections.abc import Iterator def _UpperCamelCase ( UpperCamelCase = "." ) -> Iterator[str]: """simple docstring""" for dir_path, dir_names, filenames in os.walk(UpperCamelCase ): __UpperCAmelCase : Any = [d for d in dir_names if d != "scripts" and d[0] not in "._"] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(UpperCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(UpperCamelCase , UpperCamelCase ).lstrip("./" ) def _UpperCamelCase ( UpperCamelCase ) -> Optional[int]: """simple docstring""" return f"{i * ' '}*" if i else "\n##" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : Optional[int] = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(UpperCamelCase ) or old_parts[i] != new_part) and new_part: print(f"{md_prefix(UpperCamelCase )} {new_part.replace('_' , ' ' ).title()}" ) return new_path def _UpperCamelCase ( UpperCamelCase = "." ) -> None: """simple docstring""" __UpperCAmelCase : str = "" for filepath in sorted(good_file_paths(UpperCamelCase ) ): __UpperCAmelCase , __UpperCAmelCase : List[str] = os.path.split(UpperCamelCase ) if filepath != old_path: __UpperCAmelCase : Tuple = print_path(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = (filepath.count(os.sep ) + 1) if filepath else 0 __UpperCAmelCase : Any = f"{filepath}/{filename}".replace(" " , "%20" ) __UpperCAmelCase : int = os.path.splitext(filename.replace("_" , " " ).title() )[0] print(f"{md_prefix(UpperCamelCase )} [{filename}]({url})" ) if __name__ == "__main__": print_directory_md(""".""")
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) A = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { """microsoft/unispeech-sat-base-100h-libri-ft""": ( """https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json""" ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class a__ ( __magic_name__ ): lowercase_ = "unispeech-sat" def __init__( self : str , UpperCamelCase_ : List[str]=32 , UpperCamelCase_ : Tuple=768 , UpperCamelCase_ : str=12 , UpperCamelCase_ : Optional[int]=12 , UpperCamelCase_ : List[str]=3072 , UpperCamelCase_ : Optional[Any]="gelu" , UpperCamelCase_ : str=0.1 , UpperCamelCase_ : int=0.1 , UpperCamelCase_ : Optional[int]=0.1 , UpperCamelCase_ : Dict=0.0 , UpperCamelCase_ : Tuple=0.0 , UpperCamelCase_ : Dict=0.1 , UpperCamelCase_ : Any=0.1 , UpperCamelCase_ : Optional[int]=0.02 , UpperCamelCase_ : Tuple=1e-5 , UpperCamelCase_ : List[str]="group" , UpperCamelCase_ : Optional[Any]="gelu" , UpperCamelCase_ : Union[str, Any]=(512, 512, 512, 512, 512, 512, 512) , UpperCamelCase_ : Dict=(5, 2, 2, 2, 2, 2, 2) , UpperCamelCase_ : str=(10, 3, 3, 3, 3, 2, 2) , UpperCamelCase_ : Union[str, Any]=False , UpperCamelCase_ : Union[str, Any]=128 , UpperCamelCase_ : Tuple=16 , UpperCamelCase_ : Tuple=False , UpperCamelCase_ : List[str]=True , UpperCamelCase_ : Optional[Any]=0.05 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : int=2 , UpperCamelCase_ : List[Any]=0.0 , UpperCamelCase_ : List[Any]=10 , UpperCamelCase_ : Dict=0 , UpperCamelCase_ : Tuple=320 , UpperCamelCase_ : List[Any]=2 , UpperCamelCase_ : List[Any]=0.1 , UpperCamelCase_ : List[Any]=100 , UpperCamelCase_ : List[str]=256 , UpperCamelCase_ : Optional[int]=256 , UpperCamelCase_ : Dict=0.1 , UpperCamelCase_ : Union[str, Any]="mean" , UpperCamelCase_ : str=False , UpperCamelCase_ : Dict=False , UpperCamelCase_ : Optional[Any]=256 , UpperCamelCase_ : int=(512, 512, 512, 512, 1500) , UpperCamelCase_ : Tuple=(5, 3, 3, 1, 1) , UpperCamelCase_ : Optional[Any]=(1, 2, 3, 1, 1) , UpperCamelCase_ : int=512 , UpperCamelCase_ : str=0 , UpperCamelCase_ : str=1 , UpperCamelCase_ : Optional[int]=2 , UpperCamelCase_ : List[Any]=504 , **UpperCamelCase_ : str , ): """simple docstring""" super().__init__(**UpperCamelCase_ , pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_) __UpperCAmelCase : Tuple = hidden_size __UpperCAmelCase : Optional[Any] = feat_extract_norm __UpperCAmelCase : List[str] = feat_extract_activation __UpperCAmelCase : Optional[int] = list(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = list(UpperCamelCase_) __UpperCAmelCase : List[str] = list(UpperCamelCase_) __UpperCAmelCase : str = conv_bias __UpperCAmelCase : Tuple = num_conv_pos_embeddings __UpperCAmelCase : str = num_conv_pos_embedding_groups __UpperCAmelCase : List[str] = len(self.conv_dim) __UpperCAmelCase : Tuple = num_hidden_layers __UpperCAmelCase : Tuple = intermediate_size __UpperCAmelCase : Any = hidden_act __UpperCAmelCase : Optional[Any] = num_attention_heads __UpperCAmelCase : Any = hidden_dropout __UpperCAmelCase : Tuple = attention_dropout __UpperCAmelCase : Optional[int] = activation_dropout __UpperCAmelCase : Union[str, Any] = feat_proj_dropout __UpperCAmelCase : str = final_dropout __UpperCAmelCase : Tuple = layerdrop __UpperCAmelCase : List[Any] = layer_norm_eps __UpperCAmelCase : Any = initializer_range __UpperCAmelCase : List[Any] = vocab_size __UpperCAmelCase : Tuple = num_clusters __UpperCAmelCase : Dict = do_stable_layer_norm __UpperCAmelCase : List[Any] = use_weighted_layer_sum if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" F" {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`," F" `len(config.conv_kernel) = {len(self.conv_kernel)}`.") # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCAmelCase : Dict = apply_spec_augment __UpperCAmelCase : Union[str, Any] = mask_time_prob __UpperCAmelCase : Optional[Any] = mask_time_length __UpperCAmelCase : Tuple = mask_time_min_masks __UpperCAmelCase : int = mask_feature_prob __UpperCAmelCase : Optional[Any] = mask_feature_length __UpperCAmelCase : Any = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __UpperCAmelCase : int = num_codevectors_per_group __UpperCAmelCase : str = num_codevector_groups __UpperCAmelCase : Any = contrastive_logits_temperature __UpperCAmelCase : List[str] = feat_quantizer_dropout __UpperCAmelCase : List[Any] = num_negatives __UpperCAmelCase : List[Any] = codevector_dim __UpperCAmelCase : Tuple = proj_codevector_dim __UpperCAmelCase : Tuple = diversity_loss_weight # ctc loss __UpperCAmelCase : int = ctc_loss_reduction __UpperCAmelCase : Any = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. __UpperCAmelCase : Optional[Any] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __UpperCAmelCase : List[Any] = list(UpperCamelCase_) __UpperCAmelCase : List[str] = list(UpperCamelCase_) __UpperCAmelCase : Any = list(UpperCamelCase_) __UpperCAmelCase : Tuple = xvector_output_dim @property def a_ ( self : Optional[int]): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1)
77
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCamelCase_ : int , UpperCamelCase_ : Any=13 , UpperCamelCase_ : Optional[int]=3 , UpperCamelCase_ : int=224 , UpperCamelCase_ : int=30 , UpperCamelCase_ : str=400 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Optional[int]=[0.5, 0.5, 0.5] , UpperCamelCase_ : Optional[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __UpperCAmelCase : Tuple = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : List[Any] = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : List[Any] = image_size __UpperCAmelCase : str = min_resolution __UpperCAmelCase : Tuple = max_resolution __UpperCAmelCase : Optional[Any] = do_resize __UpperCAmelCase : Any = size __UpperCAmelCase : Any = do_normalize __UpperCAmelCase : Any = image_mean __UpperCAmelCase : Optional[Any] = image_std def a_ ( self : str): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ViTImageProcessor if is_vision_available() else None def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Optional[Any] = EfficientFormerImageProcessorTester(self) @property def a_ ( self : Union[str, Any]): """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCamelCase_ , "image_std")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) def a_ ( self : Dict): """simple docstring""" pass def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : str = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Optional[int] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : Tuple = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Any = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
77
1
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() A = logging.get_logger(__name__) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" __UpperCAmelCase : Any = WavaVecaForSequenceClassification.from_pretrained(UpperCamelCase , config=UpperCamelCase ) __UpperCAmelCase : int = downstream_dict["projector.weight"] __UpperCAmelCase : List[Any] = downstream_dict["projector.bias"] __UpperCAmelCase : Optional[Any] = downstream_dict["model.post_net.linear.weight"] __UpperCAmelCase : List[Any] = downstream_dict["model.post_net.linear.bias"] return model def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Union[str, Any] = WavaVecaForAudioFrameClassification.from_pretrained(UpperCamelCase , config=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = downstream_dict["model.linear.weight"] __UpperCAmelCase : Union[str, Any] = downstream_dict["model.linear.bias"] return model def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : List[str] = WavaVecaForXVector.from_pretrained(UpperCamelCase , config=UpperCamelCase ) __UpperCAmelCase : Tuple = downstream_dict["connector.weight"] __UpperCAmelCase : str = downstream_dict["connector.bias"] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): __UpperCAmelCase : int = downstream_dict[ f"model.framelevel_feature_extractor.module.{i}.kernel.weight" ] __UpperCAmelCase : Union[str, Any] = downstream_dict[f"model.framelevel_feature_extractor.module.{i}.kernel.bias"] __UpperCAmelCase : Tuple = downstream_dict["model.utterancelevel_feature_extractor.linear1.weight"] __UpperCAmelCase : Union[str, Any] = downstream_dict["model.utterancelevel_feature_extractor.linear1.bias"] __UpperCAmelCase : Dict = downstream_dict["model.utterancelevel_feature_extractor.linear2.weight"] __UpperCAmelCase : Union[str, Any] = downstream_dict["model.utterancelevel_feature_extractor.linear2.bias"] __UpperCAmelCase : int = downstream_dict["objective.W"] return model @torch.no_grad() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Dict: """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.load(UpperCamelCase , map_location="cpu" ) __UpperCAmelCase : Optional[Any] = checkpoint["Downstream"] __UpperCAmelCase : int = WavaVecaConfig.from_pretrained(UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = WavaVecaFeatureExtractor.from_pretrained( UpperCamelCase , return_attention_mask=UpperCamelCase , do_normalize=UpperCamelCase ) __UpperCAmelCase : Dict = hf_config.architectures[0] if arch.endswith("ForSequenceClassification" ): __UpperCAmelCase : List[Any] = convert_classification(UpperCamelCase , UpperCamelCase , UpperCamelCase ) elif arch.endswith("ForAudioFrameClassification" ): __UpperCAmelCase : List[str] = convert_diarization(UpperCamelCase , UpperCamelCase , UpperCamelCase ) elif arch.endswith("ForXVector" ): __UpperCAmelCase : str = convert_xvector(UpperCamelCase , UpperCamelCase , UpperCamelCase ) else: raise NotImplementedError(f"S3PRL weights conversion is not supported for {arch}" ) if hf_config.use_weighted_layer_sum: __UpperCAmelCase : Optional[Any] = checkpoint["Featurizer"]["weights"] hf_feature_extractor.save_pretrained(UpperCamelCase ) hf_model.save_pretrained(UpperCamelCase ) if __name__ == "__main__": A = argparse.ArgumentParser() parser.add_argument( """--base_model_name""", default=None, type=str, help="""Name of the huggingface pretrained base model.""" ) parser.add_argument("""--config_path""", default=None, type=str, help="""Path to the huggingface classifier config.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to the s3prl checkpoint.""") parser.add_argument("""--model_dump_path""", default=None, type=str, help="""Path to the final converted model.""") A = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
77
"""simple docstring""" from collections import namedtuple A = namedtuple("""from_to""", """from_ to""") A = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1_000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": from_to(0.76455, 1.30795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.000236588, 4226.75), } def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'from_type' value: {from_type!r} Supported values are:\n" + ", ".join(UpperCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n" + ", ".join(UpperCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
77
1
"""simple docstring""" from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging A = logging.get_logger(__name__) A = { """Salesforce/codegen-350M-nl""": """https://huggingface.co/Salesforce/codegen-350M-nl/resolve/main/config.json""", """Salesforce/codegen-350M-multi""": """https://huggingface.co/Salesforce/codegen-350M-multi/resolve/main/config.json""", """Salesforce/codegen-350M-mono""": """https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/config.json""", """Salesforce/codegen-2B-nl""": """https://huggingface.co/Salesforce/codegen-2B-nl/resolve/main/config.json""", """Salesforce/codegen-2B-multi""": """https://huggingface.co/Salesforce/codegen-2B-multi/resolve/main/config.json""", """Salesforce/codegen-2B-mono""": """https://huggingface.co/Salesforce/codegen-2B-mono/resolve/main/config.json""", """Salesforce/codegen-6B-nl""": """https://huggingface.co/Salesforce/codegen-6B-nl/resolve/main/config.json""", """Salesforce/codegen-6B-multi""": """https://huggingface.co/Salesforce/codegen-6B-multi/resolve/main/config.json""", """Salesforce/codegen-6B-mono""": """https://huggingface.co/Salesforce/codegen-6B-mono/resolve/main/config.json""", """Salesforce/codegen-16B-nl""": """https://huggingface.co/Salesforce/codegen-16B-nl/resolve/main/config.json""", """Salesforce/codegen-16B-multi""": """https://huggingface.co/Salesforce/codegen-16B-multi/resolve/main/config.json""", """Salesforce/codegen-16B-mono""": """https://huggingface.co/Salesforce/codegen-16B-mono/resolve/main/config.json""", } class a__ ( __magic_name__ ): lowercase_ = "codegen" lowercase_ = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Optional[int] , UpperCamelCase_ : Optional[int]=50400 , UpperCamelCase_ : Tuple=2048 , UpperCamelCase_ : Optional[int]=2048 , UpperCamelCase_ : Optional[int]=4096 , UpperCamelCase_ : Optional[Any]=28 , UpperCamelCase_ : Union[str, Any]=16 , UpperCamelCase_ : List[Any]=64 , UpperCamelCase_ : Dict=None , UpperCamelCase_ : int="gelu_new" , UpperCamelCase_ : Any=0.0 , UpperCamelCase_ : str=0.0 , UpperCamelCase_ : Any=0.0 , UpperCamelCase_ : List[str]=1e-5 , UpperCamelCase_ : Union[str, Any]=0.02 , UpperCamelCase_ : Union[str, Any]=True , UpperCamelCase_ : List[str]=50256 , UpperCamelCase_ : int=50256 , UpperCamelCase_ : Optional[Any]=False , **UpperCamelCase_ : List[str] , ): """simple docstring""" __UpperCAmelCase : Tuple = vocab_size __UpperCAmelCase : Union[str, Any] = n_ctx __UpperCAmelCase : List[str] = n_positions __UpperCAmelCase : Dict = n_embd __UpperCAmelCase : List[Any] = n_layer __UpperCAmelCase : List[Any] = n_head __UpperCAmelCase : int = n_inner __UpperCAmelCase : Optional[int] = rotary_dim __UpperCAmelCase : Union[str, Any] = activation_function __UpperCAmelCase : Optional[Any] = resid_pdrop __UpperCAmelCase : Tuple = embd_pdrop __UpperCAmelCase : List[Any] = attn_pdrop __UpperCAmelCase : Union[str, Any] = layer_norm_epsilon __UpperCAmelCase : Tuple = initializer_range __UpperCAmelCase : Optional[Any] = use_cache __UpperCAmelCase : Union[str, Any] = bos_token_id __UpperCAmelCase : Optional[Any] = eos_token_id super().__init__( bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , tie_word_embeddings=UpperCamelCase_ , **UpperCamelCase_) class a__ ( __magic_name__ ): def __init__( self : List[str] , UpperCamelCase_ : PretrainedConfig , UpperCamelCase_ : str = "default" , UpperCamelCase_ : List[PatchingSpec] = None , UpperCamelCase_ : bool = False , ): """simple docstring""" super().__init__(UpperCamelCase_ , task=UpperCamelCase_ , patching_specs=UpperCamelCase_ , use_past=UpperCamelCase_) if not getattr(self._config , "pad_token_id" , UpperCamelCase_): # TODO: how to do that better? __UpperCAmelCase : Optional[Any] = 0 @property def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}}) if self.use_past: self.fill_with_past_key_values_(UpperCamelCase_ , direction="inputs") __UpperCAmelCase : Dict = {0: "batch", 1: "past_sequence + sequence"} else: __UpperCAmelCase : int = {0: "batch", 1: "sequence"} return common_inputs @property def a_ ( self : Optional[Any]): """simple docstring""" return self._config.n_layer @property def a_ ( self : Optional[int]): """simple docstring""" return self._config.n_head def a_ ( self : Optional[int] , UpperCamelCase_ : PreTrainedTokenizer , UpperCamelCase_ : int = -1 , UpperCamelCase_ : int = -1 , UpperCamelCase_ : bool = False , UpperCamelCase_ : Optional[TensorType] = None , ): """simple docstring""" __UpperCAmelCase : str = super(UpperCamelCase_ , self).generate_dummy_inputs( UpperCamelCase_ , batch_size=UpperCamelCase_ , seq_length=UpperCamelCase_ , is_pair=UpperCamelCase_ , framework=UpperCamelCase_) # We need to order the input in the way they appears in the forward() __UpperCAmelCase : str = OrderedDict({"input_ids": common_inputs["input_ids"]}) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed.") else: import torch __UpperCAmelCase , __UpperCAmelCase : Dict = common_inputs["input_ids"].shape # Not using the same length for past_key_values __UpperCAmelCase : List[Any] = seqlen + 2 __UpperCAmelCase : List[Any] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __UpperCAmelCase : Any = [ (torch.zeros(UpperCamelCase_), torch.zeros(UpperCamelCase_)) for _ in range(self.num_layers) ] __UpperCAmelCase : Dict = common_inputs["attention_mask"] if self.use_past: __UpperCAmelCase : Any = ordered_inputs["attention_mask"].dtype __UpperCAmelCase : Any = torch.cat( [ordered_inputs["attention_mask"], torch.ones(UpperCamelCase_ , UpperCamelCase_ , dtype=UpperCamelCase_)] , dim=1) return ordered_inputs @property def a_ ( self : Union[str, Any]): """simple docstring""" return 13
77
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ShapEPipeline lowercase_ = ["prompt"] lowercase_ = ["prompt"] lowercase_ = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] lowercase_ = False @property def a_ ( self : Optional[int]): """simple docstring""" return 32 @property def a_ ( self : Any): """simple docstring""" return 32 @property def a_ ( self : int): """simple docstring""" return self.time_input_dim * 4 @property def a_ ( self : List[Any]): """simple docstring""" return 8 @property def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") return tokenizer @property def a_ ( self : List[str]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(UpperCamelCase_) @property def a_ ( self : Any): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Union[str, Any] = { "num_attention_heads": 2, "attention_head_dim": 16, "embedding_dim": self.time_input_dim, "num_embeddings": 32, "embedding_proj_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "num_layers": 1, "clip_embed_dim": self.time_input_dim * 2, "additional_embeddings": 0, "time_embed_act_fn": "gelu", "norm_in_type": "layer", "encoder_hid_proj_type": None, "added_emb_type": None, } __UpperCAmelCase : Dict = PriorTransformer(**UpperCamelCase_) return model @property def a_ ( self : Union[str, Any]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Tuple = { "param_shapes": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), "d_latent": self.time_input_dim, "d_hidden": self.renderer_dim, "n_output": 12, "background": ( 0.1, 0.1, 0.1, ), } __UpperCAmelCase : List[Any] = ShapERenderer(**UpperCamelCase_) return model def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.dummy_prior __UpperCAmelCase : str = self.dummy_text_encoder __UpperCAmelCase : int = self.dummy_tokenizer __UpperCAmelCase : int = self.dummy_renderer __UpperCAmelCase : Tuple = HeunDiscreteScheduler( beta_schedule="exp" , num_train_timesteps=1024 , prediction_type="sample" , use_karras_sigmas=UpperCamelCase_ , clip_sample=UpperCamelCase_ , clip_sample_range=1.0 , ) __UpperCAmelCase : str = { "prior": prior, "text_encoder": text_encoder, "tokenizer": tokenizer, "renderer": renderer, "scheduler": scheduler, } return components def a_ ( self : Optional[int] , UpperCamelCase_ : Dict , UpperCamelCase_ : Any=0): """simple docstring""" if str(UpperCamelCase_).startswith("mps"): __UpperCAmelCase : List[Any] = torch.manual_seed(UpperCamelCase_) else: __UpperCAmelCase : str = torch.Generator(device=UpperCamelCase_).manual_seed(UpperCamelCase_) __UpperCAmelCase : List[Any] = { "prompt": "horse", "generator": generator, "num_inference_steps": 1, "frame_size": 32, "output_type": "np", } return inputs def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : str = "cpu" __UpperCAmelCase : Union[str, Any] = self.get_dummy_components() __UpperCAmelCase : Union[str, Any] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[Any] = pipe(**self.get_dummy_inputs(UpperCamelCase_)) __UpperCAmelCase : Union[str, Any] = output.images[0] __UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __UpperCAmelCase : Union[str, Any] = np.array( [ 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def a_ ( self : Tuple): """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2]) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch_device == "cpu" __UpperCAmelCase : Optional[Any] = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=UpperCamelCase_ , relax_max_difference=UpperCamelCase_ , ) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.get_dummy_components() __UpperCAmelCase : List[str] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : int = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[int] = 1 __UpperCAmelCase : Any = 2 __UpperCAmelCase : Optional[Any] = self.get_dummy_inputs(UpperCamelCase_) for key in inputs.keys(): if key in self.batch_params: __UpperCAmelCase : List[Any] = batch_size * [inputs[key]] __UpperCAmelCase : List[Any] = pipe(**UpperCamelCase_ , num_images_per_prompt=UpperCamelCase_)[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class a__ ( unittest.TestCase ): def a_ ( self : List[str]): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/test_shap_e_np_out.npy") __UpperCAmelCase : Optional[Any] = ShapEPipeline.from_pretrained("openai/shap-e") __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Dict = torch.Generator(device=UpperCamelCase_).manual_seed(0) __UpperCAmelCase : int = pipe( "a shark" , generator=UpperCamelCase_ , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type="np" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(UpperCamelCase_ , UpperCamelCase_)
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = 0 ) -> list: """simple docstring""" __UpperCAmelCase : List[str] = length or len(UpperCamelCase ) __UpperCAmelCase : List[str] = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: __UpperCAmelCase , __UpperCAmelCase : int = list_data[i + 1], list_data[i] __UpperCAmelCase : List[Any] = True return list_data if not swapped else bubble_sort(UpperCamelCase , length - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_features", "is_longer"] def __init__( self : List[str] , UpperCamelCase_ : Dict=64 , UpperCamelCase_ : Tuple=48000 , UpperCamelCase_ : List[Any]=480 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=1024 , UpperCamelCase_ : List[str]=0.0 , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : float = 0 , UpperCamelCase_ : float = 14000 , UpperCamelCase_ : int = None , UpperCamelCase_ : str = "fusion" , UpperCamelCase_ : str = "repeatpad" , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" super().__init__( feature_size=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , padding_value=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = top_db __UpperCAmelCase : Optional[Any] = truncation __UpperCAmelCase : str = padding __UpperCAmelCase : int = fft_window_size __UpperCAmelCase : str = (fft_window_size >> 1) + 1 __UpperCAmelCase : List[Any] = hop_length __UpperCAmelCase : Optional[Any] = max_length_s __UpperCAmelCase : Tuple = max_length_s * sampling_rate __UpperCAmelCase : str = sampling_rate __UpperCAmelCase : int = frequency_min __UpperCAmelCase : Optional[Any] = frequency_max __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm=UpperCamelCase_ , mel_scale="htk" , ) __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm="slaney" , mel_scale="slaney" , ) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Dict = copy.deepcopy(self.__dict__) __UpperCAmelCase : str = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def a_ ( self : int , UpperCamelCase_ : np.array , UpperCamelCase_ : Optional[np.array] = None): """simple docstring""" __UpperCAmelCase : List[Any] = spectrogram( UpperCamelCase_ , window_function(self.fft_window_size , "hann") , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=UpperCamelCase_ , log_mel="dB" , ) return log_mel_spectrogram.T def a_ ( self : List[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = np.array_split(list(range(0 , total_frames - chunk_frames + 1)) , 3) if len(ranges[1]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : str = [0] if len(ranges[2]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : Dict = [0] # randomly choose index for each part __UpperCAmelCase : Dict = np.random.choice(ranges[0]) __UpperCAmelCase : List[str] = np.random.choice(ranges[1]) __UpperCAmelCase : List[Any] = np.random.choice(ranges[2]) __UpperCAmelCase : List[Any] = mel[idx_front : idx_front + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_back : idx_back + chunk_frames, :] __UpperCAmelCase : Tuple = torch.tensor(mel[None, None, :]) __UpperCAmelCase : Union[str, Any] = torch.nn.functional.interpolate( UpperCamelCase_ , size=[chunk_frames, 64] , mode="bilinear" , align_corners=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = mel_shrink[0][0].numpy() __UpperCAmelCase : Optional[int] = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0) return mel_fusion def a_ ( self : Optional[Any] , UpperCamelCase_ : np.array , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[Any]): """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": __UpperCAmelCase : List[str] = True # random crop to max_length (for compatibility) -> this should be handled by self.pad __UpperCAmelCase : List[Any] = len(UpperCamelCase_) - max_length __UpperCAmelCase : int = np.random.randint(0 , overflow + 1) __UpperCAmelCase : Union[str, Any] = waveform[idx : idx + max_length] __UpperCAmelCase : Union[str, Any] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] elif truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : Dict = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed __UpperCAmelCase : Tuple = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. __UpperCAmelCase : List[str] = np.stack([mel, mel, mel, mel] , axis=0) __UpperCAmelCase : Any = False else: __UpperCAmelCase : List[str] = self._random_mel_fusion(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented") else: __UpperCAmelCase : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": __UpperCAmelCase : Tuple = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : List[str] = np.stack(np.tile(UpperCamelCase_ , n_repeat + 1))[:max_length] if padding == "repeatpad": __UpperCAmelCase : Union[str, Any] = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : Optional[Any] = np.stack(np.tile(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : int = np.pad(UpperCamelCase_ , (0, max_length - waveform.shape[0]) , mode="constant" , constant_values=0) if truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0) else: __UpperCAmelCase : Optional[int] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] return input_mel, longer def __call__( self : Dict , UpperCamelCase_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase_ : str = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[Union[str, TensorType]] = None , **UpperCamelCase_ : Any , ): """simple docstring""" __UpperCAmelCase : int = truncation if truncation is not None else self.truncation __UpperCAmelCase : Optional[Any] = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}.") else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug.") __UpperCAmelCase : List[str] = isinstance(UpperCamelCase_ , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}") __UpperCAmelCase : str = is_batched_numpy or ( isinstance(UpperCamelCase_ , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: __UpperCAmelCase : Dict = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase_ , np.ndarray): __UpperCAmelCase : Tuple = np.asarray(UpperCamelCase_ , dtype=np.floataa) elif isinstance(UpperCamelCase_ , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): __UpperCAmelCase : Optional[int] = raw_speech.astype(np.floataa) # always return batch if not is_batched: __UpperCAmelCase : int = [np.asarray(UpperCamelCase_)] # convert to mel spectrogram, truncate and pad if needed. __UpperCAmelCase : Optional[int] = [ self._get_input_mel(UpperCamelCase_ , max_length if max_length else self.nb_max_samples , UpperCamelCase_ , UpperCamelCase_) for waveform in raw_speech ] __UpperCAmelCase : Tuple = [] __UpperCAmelCase : List[Any] = [] for mel, longer in padded_inputs: input_mel.append(UpperCamelCase_) is_longer.append(UpperCamelCase_) if truncation == "fusion" and sum(UpperCamelCase_) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer __UpperCAmelCase : Any = np.random.randint(0 , len(UpperCamelCase_)) __UpperCAmelCase : Optional[int] = True if isinstance(input_mel[0] , UpperCamelCase_): __UpperCAmelCase : Tuple = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for feature in input_mel] # is_longer is a list of bool __UpperCAmelCase : List[str] = [[longer] for longer in is_longer] __UpperCAmelCase : Optional[int] = {"input_features": input_mel, "is_longer": is_longer} __UpperCAmelCase : Optional[int] = BatchFeature(UpperCamelCase_) if return_tensors is not None: __UpperCAmelCase : Any = input_features.convert_to_tensors(UpperCamelCase_) return input_features
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { """configuration_mobilebert""": [ """MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileBertConfig""", """MobileBertOnnxConfig""", ], """tokenization_mobilebert""": ["""MobileBertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""MobileBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileBertForMaskedLM""", """MobileBertForMultipleChoice""", """MobileBertForNextSentencePrediction""", """MobileBertForPreTraining""", """MobileBertForQuestionAnswering""", """MobileBertForSequenceClassification""", """MobileBertForTokenClassification""", """MobileBertLayer""", """MobileBertModel""", """MobileBertPreTrainedModel""", """load_tf_weights_in_mobilebert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileBertForMaskedLM""", """TFMobileBertForMultipleChoice""", """TFMobileBertForNextSentencePrediction""", """TFMobileBertForPreTraining""", """TFMobileBertForQuestionAnswering""", """TFMobileBertForSequenceClassification""", """TFMobileBertForTokenClassification""", """TFMobileBertMainLayer""", """TFMobileBertModel""", """TFMobileBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , UpperCamelCase_ : List[Any]="</s>" , UpperCamelCase_ : Tuple="<unk>" , UpperCamelCase_ : List[str]="<pad>" , UpperCamelCase_ : Union[str, Any]=125 , UpperCamelCase_ : Dict=None , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: __UpperCAmelCase : int = [F"<extra_id_{i}>" for i in range(UpperCamelCase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens __UpperCAmelCase : Dict = len(set(filter(lambda UpperCamelCase_: bool("extra_id" in str(UpperCamelCase_)) , UpperCamelCase_))) if extra_tokens != extra_ids: raise ValueError( F"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" " provided to ByT5Tokenizer. In this case the additional_special_tokens must include the" " extra_ids tokens") __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else pad_token __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else eos_token __UpperCAmelCase : Optional[int] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else unk_token super().__init__( eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , extra_ids=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : List[str] = extra_ids __UpperCAmelCase : int = 2**8 # utf is 8 bits # define special tokens dict __UpperCAmelCase : Dict[int, str] = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } __UpperCAmelCase : Any = len(self.special_tokens_encoder) __UpperCAmelCase : List[Any] = len(UpperCamelCase_) for i, token in enumerate(UpperCamelCase_): __UpperCAmelCase : Union[str, Any] = self.vocab_size + i - n __UpperCAmelCase : Dict[str, int] = {v: k for k, v in self.special_tokens_encoder.items()} @property def a_ ( self : List[Any]): """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def a_ ( self : List[str] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(UpperCamelCase_)) + [1] return ([0] * len(UpperCamelCase_)) + [1] + ([0] * len(UpperCamelCase_)) + [1] def a_ ( self : Optional[Any] , UpperCamelCase_ : List[int]): """simple docstring""" if len(UpperCamelCase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F"This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated" " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def a_ ( self : Dict , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Dict = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def a_ ( self : Optional[int] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Optional[Any] = self._add_eos_if_not_present(UpperCamelCase_) if token_ids_a is None: return token_ids_a else: __UpperCAmelCase : List[Any] = self._add_eos_if_not_present(UpperCamelCase_) return token_ids_a + token_ids_a def a_ ( self : List[str] , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : Any = [chr(UpperCamelCase_) for i in text.encode("utf-8")] return tokens def a_ ( self : Tuple , UpperCamelCase_ : List[Any]): """simple docstring""" if token in self.special_tokens_encoder: __UpperCAmelCase : Any = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: __UpperCAmelCase : int = self.added_tokens_encoder[token] elif len(UpperCamelCase_) != 1: __UpperCAmelCase : Optional[Any] = self.unk_token_id else: __UpperCAmelCase : Any = ord(UpperCamelCase_) + self._num_special_tokens return token_id def a_ ( self : Any , UpperCamelCase_ : List[str]): """simple docstring""" if index in self.special_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[index] else: __UpperCAmelCase : List[str] = chr(index - self._num_special_tokens) return token def a_ ( self : Dict , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : str = b"" for token in tokens: if token in self.special_tokens_decoder: __UpperCAmelCase : Tuple = self.special_tokens_decoder[token].encode("utf-8") elif token in self.added_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[token].encode("utf-8") elif token in self.special_tokens_encoder: __UpperCAmelCase : Optional[int] = token.encode("utf-8") elif token in self.added_tokens_encoder: __UpperCAmelCase : Optional[Any] = token.encode("utf-8") else: __UpperCAmelCase : Any = bytes([ord(UpperCamelCase_)]) bstring += tok_string __UpperCAmelCase : List[Any] = bstring.decode("utf-8" , errors="ignore") return string def a_ ( self : Optional[int] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" return ()
77
1
"""simple docstring""" # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) A = """pytorch_model.bin""" A = """pytorch_model.bin.index.json""" A = """adapter_config.json""" A = """adapter_model.bin""" A = """adapter_model.safetensors""" A = """tf_model.h5""" A = """tf_model.h5.index.json""" A = """model.ckpt""" A = """flax_model.msgpack""" A = """flax_model.msgpack.index.json""" A = """model.safetensors""" A = """model.safetensors.index.json""" A = """config.json""" A = """preprocessor_config.json""" A = FEATURE_EXTRACTOR_NAME A = """generation_config.json""" A = """modelcard.json""" A = """▁""" A = SENTENCEPIECE_UNDERLINE # Kept for backward compatibility A = [ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. A = [[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] A = [[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def _UpperCamelCase ( UpperCamelCase ) -> Tuple: """simple docstring""" if version.parse(UpperCamelCase ) < version.parse(UpperCamelCase ): if "dev" in min_version: __UpperCAmelCase : Union[str, Any] = ( "This example requires a source install from HuggingFace Transformers (see " "`https://huggingface.co/docs/transformers/installation#install-from-source`)," ) else: __UpperCAmelCase : int = f"This example requires a minimum version of {min_version}," error_message += f" but the version found is {__version__}.\n" raise ImportError( error_message + "Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other " "versions of HuggingFace Transformers." )
77
"""simple docstring""" import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a__ ( unittest.TestCase ): def __init__( self : List[Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple=3 , UpperCamelCase_ : Optional[int]=32 , UpperCamelCase_ : Dict=3 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=[10, 20, 30, 40] , UpperCamelCase_ : Tuple=[1, 1, 2, 1] , UpperCamelCase_ : str=True , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Dict="relu" , UpperCamelCase_ : str=3 , UpperCamelCase_ : int=None , ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : List[str] = image_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : Union[str, Any] = embeddings_size __UpperCAmelCase : Dict = hidden_sizes __UpperCAmelCase : Dict = depths __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_labels __UpperCAmelCase : Optional[int] = hidden_act __UpperCAmelCase : str = num_labels __UpperCAmelCase : Optional[int] = scope __UpperCAmelCase : Dict = len(UpperCamelCase_) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __UpperCAmelCase : Dict = self.get_config() return config, pixel_values def a_ ( self : Dict): """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def a_ ( self : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[str] = FlaxRegNetModel(config=UpperCamelCase_) __UpperCAmelCase : Dict = model(UpperCamelCase_) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a_ ( self : Any , UpperCamelCase_ : Dict , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : List[Any] = self.num_labels __UpperCAmelCase : Tuple = FlaxRegNetForImageClassification(config=UpperCamelCase_) __UpperCAmelCase : str = model(UpperCamelCase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Any = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase : Tuple = config_and_inputs __UpperCAmelCase : Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () lowercase_ = False lowercase_ = False lowercase_ = False def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = FlaxRegNetModelTester(self) __UpperCAmelCase : str = ConfigTester(self , config_class=UpperCamelCase_ , has_text_modality=UpperCamelCase_) def a_ ( self : Dict): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a_ ( self : Tuple): """simple docstring""" return def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase_) @unittest.skip(reason="RegNet does not use inputs_embeds") def a_ ( self : Union[str, Any]): """simple docstring""" pass @unittest.skip(reason="RegNet does not support input and output embeddings") def a_ ( self : Optional[int]): """simple docstring""" pass def a_ ( self : str): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : int = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[int] = inspect.signature(model.__call__) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : Any = [*signature.parameters.keys()] __UpperCAmelCase : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase_) def a_ ( self : int): """simple docstring""" def check_hidden_states_output(UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Union[str, Any]): __UpperCAmelCase : Union[str, Any] = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = model(**self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCAmelCase : str = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase_) , expected_num_stages + 1) __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : List[str] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase : Optional[int] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): __UpperCAmelCase : List[Any] = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[int] = model_class(UpperCamelCase_) @jax.jit def model_jitted(UpperCamelCase_ : int , **UpperCamelCase_ : Optional[int]): return model(pixel_values=UpperCamelCase_ , **UpperCamelCase_) with self.subTest("JIT Enabled"): __UpperCAmelCase : Optional[Any] = model_jitted(**UpperCamelCase_).to_tuple() with self.subTest("JIT Disabled"): with jax.disable_jit(): __UpperCAmelCase : Dict = model_jitted(**UpperCamelCase_).to_tuple() self.assertEqual(len(UpperCamelCase_) , len(UpperCamelCase_)) for jitted_output, output in zip(UpperCamelCase_ , UpperCamelCase_): self.assertEqual(jitted_output.shape , output.shape) def _UpperCamelCase ( ) -> Any: """simple docstring""" __UpperCAmelCase : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_flax class a__ ( unittest.TestCase ): @cached_property def a_ ( self : Optional[int]): """simple docstring""" return AutoImageProcessor.from_pretrained("facebook/regnet-y-040") if is_vision_available() else None @slow def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Any = FlaxRegNetForImageClassification.from_pretrained("facebook/regnet-y-040") __UpperCAmelCase : Dict = self.default_image_processor __UpperCAmelCase : str = prepare_img() __UpperCAmelCase : int = image_processor(images=UpperCamelCase_ , return_tensors="np") __UpperCAmelCase : Dict = model(**UpperCamelCase_) # verify the logits __UpperCAmelCase : Dict = (1, 1000) self.assertEqual(outputs.logits.shape , UpperCamelCase_) __UpperCAmelCase : Any = jnp.array([-0.4180, -1.5051, -3.4836]) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , UpperCamelCase_ , atol=1e-4))
77
1
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class a__ ( __magic_name__ ): lowercase_ = ["image_processor", "tokenizer"] lowercase_ = "ChineseCLIPImageProcessor" lowercase_ = ("BertTokenizer", "BertTokenizerFast") def __init__( self : Tuple , UpperCamelCase_ : List[str]=None , UpperCamelCase_ : Union[str, Any]=None , **UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : Union[str, Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = kwargs.pop("feature_extractor") __UpperCAmelCase : Union[str, Any] = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`.") if tokenizer is None: raise ValueError("You need to specify a `tokenizer`.") super().__init__(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : List[str] = self.image_processor def __call__( self : Any , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Dict=None , UpperCamelCase_ : Any=None , **UpperCamelCase_ : Optional[Any]): """simple docstring""" if text is None and images is None: raise ValueError("You have to specify either text or images. Both cannot be none.") if text is not None: __UpperCAmelCase : Dict = self.tokenizer(UpperCamelCase_ , return_tensors=UpperCamelCase_ , **UpperCamelCase_) if images is not None: __UpperCAmelCase : Any = self.image_processor(UpperCamelCase_ , return_tensors=UpperCamelCase_ , **UpperCamelCase_) if text is not None and images is not None: __UpperCAmelCase : Any = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**UpperCamelCase_) , tensor_type=UpperCamelCase_) def a_ ( self : List[str] , *UpperCamelCase_ : Union[str, Any] , **UpperCamelCase_ : Optional[int]): """simple docstring""" return self.tokenizer.batch_decode(*UpperCamelCase_ , **UpperCamelCase_) def a_ ( self : Dict , *UpperCamelCase_ : str , **UpperCamelCase_ : Any): """simple docstring""" return self.tokenizer.decode(*UpperCamelCase_ , **UpperCamelCase_) @property def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Any = self.tokenizer.model_input_names __UpperCAmelCase : Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names)) @property def a_ ( self : Optional[Any]): """simple docstring""" warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , UpperCamelCase_ , ) return self.image_processor_class
77
"""simple docstring""" from scipy.stats import spearmanr import datasets A = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ A = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ A = r"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def a_ ( self : Any): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float"), "references": datasets.Value("float"), }) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def a_ ( self : Union[str, Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : int=False): """simple docstring""" __UpperCAmelCase : List[str] = spearmanr(UpperCamelCase_ , UpperCamelCase_) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
77
1
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore A = """ Human: <<task>> Assistant: """ A = """huggingface-tools/default-prompts""" A = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase="run" ) -> List[str]: """simple docstring""" if prompt_or_repo_id is None: __UpperCAmelCase : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , UpperCamelCase ) is not None: return prompt_or_repo_id __UpperCAmelCase : str = cached_file( UpperCamelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(UpperCamelCase , "r" , encoding="utf-8" ) as f: return f.read()
77
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) A = {"""vocab_file""": """spiece.model"""} A = { """vocab_file""": { """bert_for_seq_generation""": ( """https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model""" ), } } A = {"""bert_for_seq_generation""": 512} class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = [] lowercase_ = ["input_ids", "attention_mask"] def __init__( self : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : List[str]="<s>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Optional[int]="<unk>" , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : List[Any]="<::::>" , UpperCamelCase_ : Optional[Dict[str, Any]] = None , **UpperCamelCase_ : List[Any] , ): """simple docstring""" __UpperCAmelCase : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) __UpperCAmelCase : Dict = vocab_file __UpperCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(UpperCamelCase_) @property def a_ ( self : List[str]): """simple docstring""" return self.sp_model.get_piece_size() def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : int = {self.convert_ids_to_tokens(UpperCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self : int): """simple docstring""" __UpperCAmelCase : Optional[int] = self.__dict__.copy() __UpperCAmelCase : List[Any] = None return state def __setstate__( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): __UpperCAmelCase : List[Any] = {} __UpperCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def a_ ( self : Any , UpperCamelCase_ : str): """simple docstring""" return self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_) def a_ ( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" return self.sp_model.piece_to_id(UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : int = self.sp_model.IdToPiece(UpperCamelCase_) return token def a_ ( self : Dict , UpperCamelCase_ : Optional[Any]): """simple docstring""" __UpperCAmelCase : int = [] __UpperCAmelCase : Tuple = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCamelCase_) + token __UpperCAmelCase : List[Any] = [] else: current_sub_tokens.append(UpperCamelCase_) out_string += self.sp_model.decode(UpperCamelCase_) return out_string.strip() def a_ ( self : List[str] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : Tuple = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , UpperCamelCase_) elif not os.path.isfile(self.vocab_file): with open(UpperCamelCase_ , "wb") as fi: __UpperCAmelCase : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A = { """configuration_mvp""": ["""MVP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MvpConfig""", """MvpOnnxConfig"""], """tokenization_mvp""": ["""MvpTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""MvpTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """MVP_PRETRAINED_MODEL_ARCHIVE_LIST""", """MvpForCausalLM""", """MvpForConditionalGeneration""", """MvpForQuestionAnswering""", """MvpForSequenceClassification""", """MvpModel""", """MvpPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed A = """true""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=16 ) -> Tuple: """simple docstring""" set_seed(42 ) __UpperCAmelCase : Dict = RegressionModel() __UpperCAmelCase : Optional[Any] = deepcopy(UpperCamelCase ) __UpperCAmelCase : Any = RegressionDataset(length=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = DataLoader(UpperCamelCase , batch_size=UpperCamelCase ) model.to(accelerator.device ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return model, ddp_model, dataloader def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=False ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) __UpperCAmelCase : Dict = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(UpperCamelCase ): __UpperCAmelCase : Dict = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=UpperCamelCase , max_length=UpperCamelCase ) return outputs with accelerator.main_process_first(): __UpperCAmelCase : str = dataset.map( UpperCamelCase , batched=UpperCamelCase , remove_columns=["idx", "sentence1", "sentence2"] , ) __UpperCAmelCase : List[Any] = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(UpperCamelCase ): if use_longest: return tokenizer.pad(UpperCamelCase , padding="longest" , return_tensors="pt" ) return tokenizer.pad(UpperCamelCase , padding="max_length" , max_length=128 , return_tensors="pt" ) return DataLoader(UpperCamelCase , shuffle=UpperCamelCase , collate_fn=UpperCamelCase , batch_size=16 ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[Any] = Accelerator(dispatch_batches=UpperCamelCase , split_batches=UpperCamelCase ) __UpperCAmelCase : int = get_dataloader(UpperCamelCase , not dispatch_batches ) __UpperCAmelCase : Any = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Dict = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Dict = [] for batch in dataloader: __UpperCAmelCase , __UpperCAmelCase : int = batch.values() with torch.no_grad(): __UpperCAmelCase : int = model(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : List[str] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = [], [] for logit, targ in logits_and_targets: logits.append(UpperCamelCase ) targs.append(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = torch.cat(UpperCamelCase ), torch.cat(UpperCamelCase ) return logits, targs def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=16 ) -> int: """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = get_basic_setup(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = generate_predictions(UpperCamelCase , UpperCamelCase , UpperCamelCase ) assert ( len(UpperCamelCase ) == num_samples ), f"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(UpperCamelCase )}" def _UpperCamelCase ( UpperCamelCase = False , UpperCamelCase = False ) -> List[str]: """simple docstring""" __UpperCAmelCase : List[str] = evaluate.load("glue" , "mrpc" ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = get_mrpc_setup(UpperCamelCase , UpperCamelCase ) # First do baseline __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = setup["no"] model.to(UpperCamelCase ) model.eval() for batch in dataloader: batch.to(UpperCamelCase ) with torch.inference_mode(): __UpperCAmelCase : List[str] = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=UpperCamelCase , references=batch["labels"] ) __UpperCAmelCase : str = metric.compute() # Then do distributed __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): __UpperCAmelCase : Any = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) __UpperCAmelCase : Union[str, Any] = batch["labels"] __UpperCAmelCase , __UpperCAmelCase : Any = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=UpperCamelCase , references=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def _UpperCamelCase ( ) -> List[Any]: """simple docstring""" __UpperCAmelCase : Dict = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(UpperCamelCase , UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __UpperCAmelCase : Union[str, Any] = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(UpperCamelCase , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) __UpperCAmelCase : Any = Accelerator() test_torch_metrics(UpperCamelCase , 512 ) accelerator.state._reset_state() def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
77
1
"""simple docstring""" import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex A = logging.getLogger(__name__) class a__ : def __init__( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[Any] = False def a_ ( self : Dict , UpperCamelCase_ : int , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : List[str]): """simple docstring""" if not self.initialized: __UpperCAmelCase : Union[str, Any] = RagRetriever( UpperCamelCase_ , question_encoder_tokenizer=UpperCamelCase_ , generator_tokenizer=UpperCamelCase_ , index=UpperCamelCase_ , init_retrieval=UpperCamelCase_ , ) __UpperCAmelCase : List[str] = True def a_ ( self : str): """simple docstring""" self.retriever.index.init_index() def a_ ( self : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : int = self.retriever._main_retrieve(UpperCamelCase_ , UpperCamelCase_) return doc_ids, retrieved_doc_embeds class a__ ( __magic_name__ ): def __init__( self : str , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Tuple=None): """simple docstring""" if index is not None and index.is_initialized() and len(UpperCamelCase_) > 0: raise ValueError( "When using Ray for distributed fine-tuning, " "you'll need to provide the paths instead, " "as the dataset and the index are loaded " "separately. More info in examples/rag/use_own_knowledge_dataset.py ") super().__init__( UpperCamelCase_ , question_encoder_tokenizer=UpperCamelCase_ , generator_tokenizer=UpperCamelCase_ , index=UpperCamelCase_ , init_retrieval=UpperCamelCase_ , ) __UpperCAmelCase : int = retrieval_workers if len(self.retrieval_workers) > 0: ray.get( [ worker.create_rag_retriever.remote(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) for worker in self.retrieval_workers ]) def a_ ( self : Optional[int]): """simple docstring""" logger.info("initializing retrieval") if len(self.retrieval_workers) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers]) else: # Non-distributed training. Load index into this same process. self.index.init_index() def a_ ( self : Tuple , UpperCamelCase_ : Tuple , UpperCamelCase_ : str): """simple docstring""" if len(self.retrieval_workers) > 0: # Select a random retrieval actor. __UpperCAmelCase : Any = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers) - 1)] __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = ray.get(random_worker.retrieve.remote(UpperCamelCase_ , UpperCamelCase_)) else: __UpperCAmelCase , __UpperCAmelCase : int = self._main_retrieve(UpperCamelCase_ , UpperCamelCase_) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(UpperCamelCase_) @classmethod def a_ ( cls : Dict , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Any=None , **UpperCamelCase_ : List[Any]): """simple docstring""" return super(UpperCamelCase_ , cls).get_tokenizers(UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_) @classmethod def a_ ( cls : Union[str, Any] , UpperCamelCase_ : List[str] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Union[str, Any]=None , **UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[int] = kwargs.pop("config" , UpperCamelCase_) or RagConfig.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : str = RagTokenizer.from_pretrained(UpperCamelCase_ , config=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = rag_tokenizer.question_encoder __UpperCAmelCase : Union[str, Any] = rag_tokenizer.generator if indexed_dataset is not None: __UpperCAmelCase : Union[str, Any] = "custom" __UpperCAmelCase : int = CustomHFIndex(config.retrieval_vector_size , UpperCamelCase_) else: __UpperCAmelCase : Optional[Any] = cls._build_index(UpperCamelCase_) return cls( UpperCamelCase_ , question_encoder_tokenizer=UpperCamelCase_ , generator_tokenizer=UpperCamelCase_ , retrieval_workers=UpperCamelCase_ , index=UpperCamelCase_ , )
77
"""simple docstring""" import math def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = 0 , UpperCamelCase = 0 ) -> list: """simple docstring""" __UpperCAmelCase : Union[str, Any] = end or len(UpperCamelCase ) for i in range(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[Any] = i __UpperCAmelCase : Any = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: __UpperCAmelCase : Dict = array[temp_index - 1] temp_index -= 1 __UpperCAmelCase : str = temp_index_value return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: # Max Heap """simple docstring""" __UpperCAmelCase : Optional[Any] = index __UpperCAmelCase : List[str] = 2 * index + 1 # Left Node __UpperCAmelCase : Union[str, Any] = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: __UpperCAmelCase : Tuple = left_index if right_index < heap_size and array[largest] < array[right_index]: __UpperCAmelCase : int = right_index if largest != index: __UpperCAmelCase , __UpperCAmelCase : List[str] = array[largest], array[index] heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" __UpperCAmelCase : List[Any] = len(UpperCamelCase ) for i in range(n // 2 , -1 , -1 ): heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) for i in range(n - 1 , 0 , -1 ): __UpperCAmelCase , __UpperCAmelCase : int = array[0], array[i] heapify(UpperCamelCase , 0 , UpperCamelCase ) return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Optional[Any] = low __UpperCAmelCase : List[str] = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i __UpperCAmelCase , __UpperCAmelCase : Optional[int] = array[j], array[i] i += 1 def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" if len(UpperCamelCase ) == 0: return array __UpperCAmelCase : Optional[int] = 2 * math.ceil(math.loga(len(UpperCamelCase ) ) ) __UpperCAmelCase : List[Any] = 16 return intro_sort(UpperCamelCase , 0 , len(UpperCamelCase ) , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> list: """simple docstring""" while end - start > size_threshold: if max_depth == 0: return heap_sort(UpperCamelCase ) max_depth -= 1 __UpperCAmelCase : List[Any] = median_of_a(UpperCamelCase , UpperCamelCase , start + ((end - start) // 2) + 1 , end - 1 ) __UpperCAmelCase : Union[str, Any] = partition(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) intro_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[Any] = p return insertion_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() A = input("""Enter numbers separated by a comma : """).strip() A = [float(item) for item in user_input.split(""",""")] print(sort(unsorted))
77
1
"""simple docstring""" # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A = { """configuration_xmod""": [ """XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XmodConfig""", """XmodOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """XMOD_PRETRAINED_MODEL_ARCHIVE_LIST""", """XmodForCausalLM""", """XmodForMaskedLM""", """XmodForMultipleChoice""", """XmodForQuestionAnswering""", """XmodForSequenceClassification""", """XmodForTokenClassification""", """XmodModel""", """XmodPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" import numpy as np from PIL import Image def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : str = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Any = 0 __UpperCAmelCase : Dict = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Tuple = 0 # compute the shape of the output matrix __UpperCAmelCase : Optional[int] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __UpperCAmelCase : List[str] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __UpperCAmelCase : str = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : int = 0 __UpperCAmelCase : int = 0 return updated_arr def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : List[str] = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Any = 0 # compute the shape of the output matrix __UpperCAmelCase : Tuple = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __UpperCAmelCase : str = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __UpperCAmelCase : Tuple = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Optional[Any] = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name="""avgpooling""", verbose=True) # Loading the image A = Image.open("""path_to_image""") # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
77
1
"""simple docstring""" import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class a__ ( unittest.TestCase ): def __init__( self : str , UpperCamelCase_ : str , UpperCamelCase_ : List[str]=13 , UpperCamelCase_ : Optional[int]=7 , UpperCamelCase_ : Union[str, Any]=True , UpperCamelCase_ : Dict=True , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Optional[Any]=True , UpperCamelCase_ : Optional[int]=99 , UpperCamelCase_ : Dict=32 , UpperCamelCase_ : int=5 , UpperCamelCase_ : int=4 , UpperCamelCase_ : List[str]=37 , UpperCamelCase_ : str="gelu" , UpperCamelCase_ : Union[str, Any]=0.1 , UpperCamelCase_ : List[str]=0.1 , UpperCamelCase_ : Dict=512 , UpperCamelCase_ : List[Any]=16 , UpperCamelCase_ : Union[str, Any]=2 , UpperCamelCase_ : Optional[Any]=0.02 , UpperCamelCase_ : Any=4 , ): """simple docstring""" __UpperCAmelCase : str = parent __UpperCAmelCase : Dict = batch_size __UpperCAmelCase : Dict = seq_length __UpperCAmelCase : List[str] = is_training __UpperCAmelCase : List[Any] = use_attention_mask __UpperCAmelCase : Union[str, Any] = use_token_type_ids __UpperCAmelCase : Dict = use_labels __UpperCAmelCase : Optional[Any] = vocab_size __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[int] = num_hidden_layers __UpperCAmelCase : List[Any] = num_attention_heads __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : Union[str, Any] = hidden_act __UpperCAmelCase : Dict = hidden_dropout_prob __UpperCAmelCase : Optional[Any] = attention_probs_dropout_prob __UpperCAmelCase : Any = max_position_embeddings __UpperCAmelCase : Any = type_vocab_size __UpperCAmelCase : Optional[int] = type_sequence_label_size __UpperCAmelCase : Any = initializer_range __UpperCAmelCase : str = num_choices def a_ ( self : int): """simple docstring""" __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __UpperCAmelCase : Union[str, Any] = None if self.use_attention_mask: __UpperCAmelCase : int = random_attention_mask([self.batch_size, self.seq_length]) __UpperCAmelCase : Dict = None if self.use_token_type_ids: __UpperCAmelCase : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __UpperCAmelCase : Union[str, Any] = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase_ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : List[Any] = config_and_inputs __UpperCAmelCase : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = config_and_inputs __UpperCAmelCase : str = True __UpperCAmelCase : int = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) __UpperCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = True lowercase_ = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def a_ ( self : str): """simple docstring""" __UpperCAmelCase : List[Any] = FlaxBertModelTester(self) @slow def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Any = FlaxBertModel.from_pretrained("bert-base-cased") __UpperCAmelCase : Union[str, Any] = model(np.ones((1, 1))) self.assertIsNotNone(UpperCamelCase_)
77
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: A = None A = logging.get_logger(__name__) A = """▁""" A = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} A = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } A = { """google/pegasus-xsum""": 512, } class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = PegasusTokenizer lowercase_ = ["input_ids", "attention_mask"] def __init__( self : str , UpperCamelCase_ : str=None , UpperCamelCase_ : Optional[int]=None , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Any="<unk>" , UpperCamelCase_ : Tuple="<mask_2>" , UpperCamelCase_ : Any="<mask_1>" , UpperCamelCase_ : Tuple=None , UpperCamelCase_ : str=103 , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" __UpperCAmelCase : Optional[int] = offset if additional_special_tokens is not None: if not isinstance(UpperCamelCase_ , UpperCamelCase_): raise TypeError( F"additional_special_tokens should be of type {type(UpperCamelCase_)}, but is" F" {type(UpperCamelCase_)}") __UpperCAmelCase : Any = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"<unk_{i}>" for i in range(len(UpperCamelCase_) , self.offset - 1) ] if len(set(UpperCamelCase_)) != len(UpperCamelCase_): raise ValueError( "Please make sure that the provided additional_special_tokens do not contain an incorrectly" F" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.") __UpperCAmelCase : str = additional_special_tokens_extended else: __UpperCAmelCase : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"<unk_{i}>" for i in range(2 , self.offset)] super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , pad_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , mask_token_sent=UpperCamelCase_ , offset=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Optional[int] = vocab_file __UpperCAmelCase : List[str] = False if not self.vocab_file else True def a_ ( self : Union[str, Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : int = set(self.all_special_ids) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens) + 3)): raise ValueError( "There should be 3 special tokens: mask_token, pad_token, and eos_token +" F" {len(self.additional_special_tokens)} additional_special_tokens, but got {all_special_ids}") return [1 if x in all_special_ids else 0 for x in seq] def a_ ( self : Union[str, Any] , UpperCamelCase_ : List , UpperCamelCase_ : Optional[List] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return self._special_token_mask(UpperCamelCase_) elif token_ids_a is None: return self._special_token_mask(UpperCamelCase_) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a) + [1] def a_ ( self : int , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : List[Any]=None): """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a_ ( self : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : List[str] = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_): copyfile(self.vocab_file , UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : List[str] = "" for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def _UpperCamelCase ( UpperCamelCase ) -> dict[str, str]: """simple docstring""" __UpperCAmelCase : int = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key __UpperCAmelCase : List[Any] = remove_duplicates(key.upper() ) __UpperCAmelCase : Dict = len(UpperCamelCase ) # First fill cipher with key characters __UpperCAmelCase : int = {alphabet[i]: char for i, char in enumerate(UpperCamelCase )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(UpperCamelCase ) , 26 ): __UpperCAmelCase : Optional[Any] = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 __UpperCAmelCase : List[Any] = alphabet[i - offset] __UpperCAmelCase : Optional[Any] = char return cipher_alphabet def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" return "".join(cipher_map.get(UpperCamelCase , UpperCamelCase ) for ch in message.upper() ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : str = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(UpperCamelCase , UpperCamelCase ) for ch in message.upper() ) def _UpperCamelCase ( ) -> None: """simple docstring""" __UpperCAmelCase : Any = input("Enter message to encode or decode: " ).strip() __UpperCAmelCase : str = input("Enter keyword: " ).strip() __UpperCAmelCase : Optional[Any] = input("Encipher or decipher? E/D:" ).strip()[0].lower() try: __UpperCAmelCase : List[str] = {"e": encipher, "d": decipher}[option] except KeyError: raise KeyError("invalid input option" ) __UpperCAmelCase : Dict = create_cipher_map(UpperCamelCase ) print(func(UpperCamelCase , UpperCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
77
"""simple docstring""" import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __UpperCAmelCase : Optional[Any] = TapasConfig.from_json_file(UpperCamelCase ) # set absolute/relative position embeddings parameter __UpperCAmelCase : Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __UpperCAmelCase : List[str] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WTQ": # run_task_main.py hparams __UpperCAmelCase : Tuple = 4 __UpperCAmelCase : Any = True # hparam_utils.py hparams __UpperCAmelCase : Union[str, Any] = 0.664694 __UpperCAmelCase : Union[str, Any] = 0.207951 __UpperCAmelCase : int = 0.121194 __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : List[str] = True __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : List[str] = 0.0352513 __UpperCAmelCase : Optional[int] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __UpperCAmelCase : int = 4 __UpperCAmelCase : Optional[int] = False # hparam_utils.py hparams __UpperCAmelCase : int = 36.4519 __UpperCAmelCase : str = 0.903421 __UpperCAmelCase : Dict = 222.088 __UpperCAmelCase : Dict = True __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Tuple = True __UpperCAmelCase : Any = 0.763141 __UpperCAmelCase : Optional[Any] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "TABFACT": __UpperCAmelCase : Union[str, Any] = TapasForSequenceClassification(config=UpperCamelCase ) elif task == "MLM": __UpperCAmelCase : Tuple = TapasForMaskedLM(config=UpperCamelCase ) elif task == "INTERMEDIATE_PRETRAINING": __UpperCAmelCase : List[str] = TapasModel(config=UpperCamelCase ) else: raise ValueError(f"Task {task} not supported." ) print(f"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # Save pytorch-model (weights and configuration) print(f"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(UpperCamelCase ) # Save tokenizer files print(f"Save tokenizer files to {pytorch_dump_path}" ) __UpperCAmelCase : str = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + "vocab.txt" , model_max_length=512 ) tokenizer.save_pretrained(UpperCamelCase ) print("Used relative position embeddings:" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
77
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import GLPNImageProcessor class a__ ( unittest.TestCase ): def __init__( self : List[str] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : List[str]=7 , UpperCamelCase_ : Union[str, Any]=3 , UpperCamelCase_ : List[Any]=18 , UpperCamelCase_ : Optional[Any]=30 , UpperCamelCase_ : List[str]=400 , UpperCamelCase_ : str=True , UpperCamelCase_ : Optional[Any]=32 , UpperCamelCase_ : Any=True , ): """simple docstring""" __UpperCAmelCase : str = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : Union[str, Any] = image_size __UpperCAmelCase : Optional[int] = min_resolution __UpperCAmelCase : Tuple = max_resolution __UpperCAmelCase : Dict = do_resize __UpperCAmelCase : List[Any] = size_divisor __UpperCAmelCase : Dict = do_rescale def a_ ( self : Union[str, Any]): """simple docstring""" return { "do_resize": self.do_resize, "size_divisor": self.size_divisor, "do_rescale": self.do_rescale, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = GLPNImageProcessor if is_vision_available() else None def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : int = GLPNImageProcessingTester(self) @property def a_ ( self : Dict): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self : str): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size_divisor")) self.assertTrue(hasattr(UpperCamelCase_ , "resample")) self.assertTrue(hasattr(UpperCamelCase_ , "do_rescale")) def a_ ( self : Union[str, Any]): """simple docstring""" pass def a_ ( self : int): """simple docstring""" __UpperCAmelCase : List[str] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input (GLPNImageProcessor doesn't support batching) __UpperCAmelCase : Any = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input (GLPNImageProcessor doesn't support batching) __UpperCAmelCase : List[str] = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Any = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input (GLPNImageProcessor doesn't support batching) __UpperCAmelCase : List[Any] = image_processing(image_inputs[0] , return_tensors="pt").pixel_values self.assertTrue(encoded_images.shape[-1] % self.image_processor_tester.size_divisor == 0) self.assertTrue(encoded_images.shape[-2] % self.image_processor_tester.size_divisor == 0)
77
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = "cpu" , UpperCamelCase = None ) -> None: """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.load(UpperCamelCase , map_location=UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(UpperCamelCase , torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) __UpperCAmelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __UpperCAmelCase : str = src_path torch.save(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
77
1
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder A = """__DUMMY_TRANSFORMERS_USER__""" A = """Dummy User""" A = """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""" A = """https://hub-ci.huggingface.co""" A = CI_HUB_ENDPOINT + """/datasets/{repo_id}/resolve/{revision}/{path}""" A = CI_HUB_ENDPOINT + """/{repo_id}/resolve/{revision}/{filename}""" A = Path("""~/.huggingface/hub_ci_token""").expanduser() @pytest.fixture def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" monkeypatch.setattr( "huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE" , UpperCamelCase ) @pytest.fixture def _UpperCamelCase ( UpperCamelCase ) -> Any: """simple docstring""" monkeypatch.setattr("datasets.config.HF_ENDPOINT" , UpperCamelCase ) monkeypatch.setattr("datasets.config.HUB_DATASETS_URL" , UpperCamelCase ) @pytest.fixture def _UpperCamelCase ( UpperCamelCase ) -> List[Any]: """simple docstring""" monkeypatch.setattr("huggingface_hub.hf_api.HfFolder.path_token" , UpperCamelCase ) @pytest.fixture def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" HfFolder.save_token(UpperCamelCase ) yield HfFolder.delete_token() @pytest.fixture(scope="session" ) def _UpperCamelCase ( ) -> Tuple: """simple docstring""" return HfApi(endpoint=UpperCamelCase ) @pytest.fixture(scope="session" ) def _UpperCamelCase ( UpperCamelCase ) -> Tuple: """simple docstring""" __UpperCAmelCase : int = HfFolder.get_token() HfFolder.save_token(UpperCamelCase ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(UpperCamelCase ) @pytest.fixture def _UpperCamelCase ( UpperCamelCase ) -> Union[str, Any]: """simple docstring""" def _cleanup_repo(UpperCamelCase ): hf_api.delete_repo(UpperCamelCase , token=UpperCamelCase , repo_type="dataset" ) return _cleanup_repo @pytest.fixture def _UpperCamelCase ( UpperCamelCase ) -> Dict: """simple docstring""" @contextmanager def _temporary_repo(UpperCamelCase ): try: yield repo_id finally: cleanup_repo(UpperCamelCase ) return _temporary_repo @pytest.fixture(scope="session" ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : str = f"repo_txt_data-{int(time.time() * 1_0e3 )}" __UpperCAmelCase : List[Any] = f"{CI_HUB_USER}/{repo_name}" hf_api.create_repo(UpperCamelCase , token=UpperCamelCase , repo_type="dataset" , private=UpperCamelCase ) hf_api.upload_file( token=UpperCamelCase , path_or_fileobj=str(UpperCamelCase ) , path_in_repo="data/text_data.txt" , repo_id=UpperCamelCase , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(UpperCamelCase , token=UpperCamelCase , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="session" ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : Dict = f"repo_zipped_txt_data-{int(time.time() * 1_0e3 )}" __UpperCAmelCase : Dict = f"{CI_HUB_USER}/{repo_name}" hf_api.create_repo(UpperCamelCase , token=UpperCamelCase , repo_type="dataset" , private=UpperCamelCase ) hf_api.upload_file( token=UpperCamelCase , path_or_fileobj=str(UpperCamelCase ) , path_in_repo="data.zip" , repo_id=UpperCamelCase , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(UpperCamelCase , token=UpperCamelCase , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Optional[Any]: """simple docstring""" return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="session" ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Any = f"repo_zipped_img_data-{int(time.time() * 1_0e3 )}" __UpperCAmelCase : List[Any] = f"{CI_HUB_USER}/{repo_name}" hf_api.create_repo(UpperCamelCase , token=UpperCamelCase , repo_type="dataset" , private=UpperCamelCase ) hf_api.upload_file( token=UpperCamelCase , path_or_fileobj=str(UpperCamelCase ) , path_in_repo="data.zip" , repo_id=UpperCamelCase , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(UpperCamelCase , token=UpperCamelCase , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Any: """simple docstring""" return hf_private_dataset_repo_zipped_img_data_
77
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": A = pd.read_csv("""sample_data.csv""", header=None) A = df.shape[:1][0] # If you're using some other dataset input the target column A = df.iloc[:, 1:2] A = actual_data.values.reshape(len_data, 1) A = MinMaxScaler().fit_transform(actual_data) A = 10 A = 5 A = 20 A = len_data - periods * look_back A = actual_data[:division] A = actual_data[division - look_back :] A , A = [], [] A , A = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) A = np.array(train_x) A = np.array(test_x) A = np.array([list(i.ravel()) for i in train_y]) A = np.array([list(i.ravel()) for i in test_y]) A = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") A = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) A = model.predict(x_test)
77
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) A = {"""vocab_file""": """spiece.model"""} A = { """vocab_file""": { """bert_for_seq_generation""": ( """https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model""" ), } } A = {"""bert_for_seq_generation""": 512} class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = [] lowercase_ = ["input_ids", "attention_mask"] def __init__( self : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : List[str]="<s>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Optional[int]="<unk>" , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : List[Any]="<::::>" , UpperCamelCase_ : Optional[Dict[str, Any]] = None , **UpperCamelCase_ : List[Any] , ): """simple docstring""" __UpperCAmelCase : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) __UpperCAmelCase : Dict = vocab_file __UpperCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(UpperCamelCase_) @property def a_ ( self : List[str]): """simple docstring""" return self.sp_model.get_piece_size() def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : int = {self.convert_ids_to_tokens(UpperCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self : int): """simple docstring""" __UpperCAmelCase : Optional[int] = self.__dict__.copy() __UpperCAmelCase : List[Any] = None return state def __setstate__( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): __UpperCAmelCase : List[Any] = {} __UpperCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def a_ ( self : Any , UpperCamelCase_ : str): """simple docstring""" return self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_) def a_ ( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" return self.sp_model.piece_to_id(UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : int = self.sp_model.IdToPiece(UpperCamelCase_) return token def a_ ( self : Dict , UpperCamelCase_ : Optional[Any]): """simple docstring""" __UpperCAmelCase : int = [] __UpperCAmelCase : Tuple = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCamelCase_) + token __UpperCAmelCase : List[Any] = [] else: current_sub_tokens.append(UpperCamelCase_) out_string += self.sp_model.decode(UpperCamelCase_) return out_string.strip() def a_ ( self : List[str] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : Tuple = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , UpperCamelCase_) elif not os.path.isfile(self.vocab_file): with open(UpperCamelCase_ , "wb") as fi: __UpperCAmelCase : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_) return (out_vocab_file,)
77
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A = 250_004 A = 250_020 @require_sentencepiece @require_tokenizers class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = MBartTokenizer lowercase_ = MBartTokenizerFast lowercase_ = True lowercase_ = True def a_ ( self : str): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase : Any = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) tokenizer.save_pretrained(self.tmpdirname) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Dict = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer.tokenize("This is a test") self.assertListEqual(UpperCamelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase_) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __UpperCAmelCase : List[Any] = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __UpperCAmelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def a_ ( self : Dict): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __UpperCAmelCase : Dict = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})"): __UpperCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = tempfile.mkdtemp() __UpperCAmelCase : Optional[int] = tokenizer_r.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Any = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) __UpperCAmelCase : Any = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f) self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Tuple = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=True __UpperCAmelCase : Optional[int] = tempfile.mkdtemp() __UpperCAmelCase : Dict = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : int = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=False __UpperCAmelCase : Tuple = tempfile.mkdtemp() __UpperCAmelCase : int = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way __UpperCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : str = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) @require_torch @require_sentencepiece @require_tokenizers class a__ ( unittest.TestCase ): lowercase_ = "facebook/mbart-large-en-ro" lowercase_ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowercase_ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowercase_ = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def a_ ( cls : int): """simple docstring""" __UpperCAmelCase : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en_XX" , tgt_lang="ro_RO") __UpperCAmelCase : Union[str, Any] = 1 return cls def a_ ( self : List[Any]): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ar_AR"] , 250001) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["en_EN"] , 250004) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ro_RO"] , 250020) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" self.assertIn(UpperCamelCase_ , self.tokenizer.all_special_ids) __UpperCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] __UpperCAmelCase : Optional[Any] = self.tokenizer.decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase_) self.assertEqual(UpperCamelCase_ , UpperCamelCase_) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , UpperCamelCase_) __UpperCAmelCase : Tuple = 10 __UpperCAmelCase : List[Any] = self.tokenizer(UpperCamelCase_ , max_length=UpperCamelCase_ , truncation=UpperCamelCase_).input_ids[0] self.assertEqual(ids[-2] , 2) self.assertEqual(ids[-1] , UpperCamelCase_) self.assertEqual(len(UpperCamelCase_) , UpperCamelCase_) def a_ ( self : Any): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"]) , [250026, 250001]) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[str] = tempfile.mkdtemp() __UpperCAmelCase : Union[str, Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase_) __UpperCAmelCase : List[Any] = MBartTokenizer.from_pretrained(UpperCamelCase_) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCamelCase_) @require_torch def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , return_tensors="pt") __UpperCAmelCase : Dict = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=len(self.expected_src_tokens) , return_tensors="pt" , ) __UpperCAmelCase : Tuple = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_) self.assertEqual((2, 14) , batch.input_ids.shape) self.assertEqual((2, 14) , batch.attention_mask.shape) __UpperCAmelCase : List[str] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) self.assertEqual(2 , batch.decoder_input_ids[0, -1]) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , []) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE]) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : List[str] = self.tokenizer(self.src_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=3 , return_tensors="pt") __UpperCAmelCase : Any = self.tokenizer( text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=10 , return_tensors="pt") __UpperCAmelCase : int = targets["input_ids"] __UpperCAmelCase : Any = shift_tokens_right(UpperCamelCase_ , self.tokenizer.pad_token_id) self.assertEqual(batch.input_ids.shape[1] , 3) self.assertEqual(batch.decoder_input_ids.shape[1] , 10) @require_torch def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="ar_AR") self.assertEqual( nested_simplify(UpperCamelCase_) , { # A, test, EOS, en_XX "input_ids": [[62, 3034, 2, 250004]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 250001, } , )
77
1
"""simple docstring""" import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed A = """true""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=16 ) -> Tuple: """simple docstring""" set_seed(42 ) __UpperCAmelCase : Dict = RegressionModel() __UpperCAmelCase : Optional[Any] = deepcopy(UpperCamelCase ) __UpperCAmelCase : Any = RegressionDataset(length=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = DataLoader(UpperCamelCase , batch_size=UpperCamelCase ) model.to(accelerator.device ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return model, ddp_model, dataloader def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=False ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) __UpperCAmelCase : Dict = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(UpperCamelCase ): __UpperCAmelCase : Dict = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=UpperCamelCase , max_length=UpperCamelCase ) return outputs with accelerator.main_process_first(): __UpperCAmelCase : str = dataset.map( UpperCamelCase , batched=UpperCamelCase , remove_columns=["idx", "sentence1", "sentence2"] , ) __UpperCAmelCase : List[Any] = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(UpperCamelCase ): if use_longest: return tokenizer.pad(UpperCamelCase , padding="longest" , return_tensors="pt" ) return tokenizer.pad(UpperCamelCase , padding="max_length" , max_length=128 , return_tensors="pt" ) return DataLoader(UpperCamelCase , shuffle=UpperCamelCase , collate_fn=UpperCamelCase , batch_size=16 ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[Any] = Accelerator(dispatch_batches=UpperCamelCase , split_batches=UpperCamelCase ) __UpperCAmelCase : int = get_dataloader(UpperCamelCase , not dispatch_batches ) __UpperCAmelCase : Any = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Dict = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Dict = [] for batch in dataloader: __UpperCAmelCase , __UpperCAmelCase : int = batch.values() with torch.no_grad(): __UpperCAmelCase : int = model(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : List[str] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = [], [] for logit, targ in logits_and_targets: logits.append(UpperCamelCase ) targs.append(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = torch.cat(UpperCamelCase ), torch.cat(UpperCamelCase ) return logits, targs def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=16 ) -> int: """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = get_basic_setup(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = generate_predictions(UpperCamelCase , UpperCamelCase , UpperCamelCase ) assert ( len(UpperCamelCase ) == num_samples ), f"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(UpperCamelCase )}" def _UpperCamelCase ( UpperCamelCase = False , UpperCamelCase = False ) -> List[str]: """simple docstring""" __UpperCAmelCase : List[str] = evaluate.load("glue" , "mrpc" ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = get_mrpc_setup(UpperCamelCase , UpperCamelCase ) # First do baseline __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = setup["no"] model.to(UpperCamelCase ) model.eval() for batch in dataloader: batch.to(UpperCamelCase ) with torch.inference_mode(): __UpperCAmelCase : List[str] = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=UpperCamelCase , references=batch["labels"] ) __UpperCAmelCase : str = metric.compute() # Then do distributed __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): __UpperCAmelCase : Any = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) __UpperCAmelCase : Union[str, Any] = batch["labels"] __UpperCAmelCase , __UpperCAmelCase : Any = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=UpperCamelCase , references=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def _UpperCamelCase ( ) -> List[Any]: """simple docstring""" __UpperCAmelCase : Dict = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(UpperCamelCase , UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __UpperCAmelCase : Union[str, Any] = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(UpperCamelCase , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) __UpperCAmelCase : Any = Accelerator() test_torch_metrics(UpperCamelCase , 512 ) accelerator.state._reset_state() def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
77
"""simple docstring""" from typing import Any class a__ : def __init__( self : List[str] , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : str = data __UpperCAmelCase : Optional[Any] = None class a__ : def __init__( self : Any): """simple docstring""" __UpperCAmelCase : Optional[int] = None def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.head while temp is not None: print(temp.data , end=" ") __UpperCAmelCase : Tuple = temp.next print() def a_ ( self : int , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : List[str] = Node(UpperCamelCase_) __UpperCAmelCase : str = self.head __UpperCAmelCase : Optional[int] = new_node def a_ ( self : str , UpperCamelCase_ : List[Any] , UpperCamelCase_ : str): """simple docstring""" if node_data_a == node_data_a: return else: __UpperCAmelCase : int = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Tuple = node_a.next __UpperCAmelCase : List[Any] = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Optional[Any] = node_a.next if node_a is None or node_a is None: return __UpperCAmelCase , __UpperCAmelCase : Any = node_a.data, node_a.data if __name__ == "__main__": A = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
77
1
"""simple docstring""" import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class a__ ( unittest.TestCase ): def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : int = 10 def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[int] = [1, 2, 3, 4] __UpperCAmelCase : str = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(UpperCamelCase_ , self.block_size , 0) , UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[Any] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] __UpperCAmelCase : Optional[int] = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(UpperCamelCase_ , self.block_size , 0) , UpperCamelCase_) def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Dict = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] __UpperCAmelCase : int = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] self.assertEqual(truncate_or_pad(UpperCamelCase_ , self.block_size , 0) , UpperCamelCase_) def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Any = "It was the year of Our Lord one thousand seven hundred and\n seventy-five.\n\nSpiritual revelations were conceded to England at that\n favoured period, as at this." __UpperCAmelCase , __UpperCAmelCase : str = process_story(UpperCamelCase_) self.assertEqual(UpperCamelCase_ , []) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : int = "" __UpperCAmelCase , __UpperCAmelCase : Tuple = process_story(UpperCamelCase_) self.assertEqual(UpperCamelCase_ , []) self.assertEqual(UpperCamelCase_ , []) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = ( "It was the year of Our Lord one thousand seven hundred and " "seventy-five\n\nSpiritual revelations were conceded to England " "at that favoured period, as at this.\n@highlight\n\nIt was the best of times" ) __UpperCAmelCase , __UpperCAmelCase : Any = process_story(UpperCamelCase_) __UpperCAmelCase : Any = [ "It was the year of Our Lord one thousand seven hundred and seventy-five.", "Spiritual revelations were conceded to England at that favoured period, as at this.", ] self.assertEqual(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : str = ["It was the best of times."] self.assertEqual(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[int] = torch.tensor([1, 2, 3, 4]) __UpperCAmelCase : List[Any] = torch.tensor([1, 1, 1, 1]) np.testing.assert_array_equal(build_mask(UpperCamelCase_ , 0).numpy() , expected.numpy()) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = torch.tensor([1, 2, 3, 4, 23, 23, 23]) __UpperCAmelCase : Dict = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(UpperCamelCase_ , 23).numpy() , expected.numpy()) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[Any] = torch.tensor([8, 2, 3, 4, 1, 1, 1]) __UpperCAmelCase : Optional[int] = torch.tensor([1, 1, 1, 1, 0, 0, 0]) np.testing.assert_array_equal(build_mask(UpperCamelCase_ , 1).numpy() , expected.numpy()) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = 101 __UpperCAmelCase : Optional[int] = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]]) __UpperCAmelCase : Tuple = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]]) __UpperCAmelCase : str = compute_token_type_ids(UpperCamelCase_ , UpperCamelCase_) np.testing.assert_array_equal(UpperCamelCase_ , UpperCamelCase_)
77
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore A = """ Human: <<task>> Assistant: """ A = """huggingface-tools/default-prompts""" A = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase="run" ) -> List[str]: """simple docstring""" if prompt_or_repo_id is None: __UpperCAmelCase : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , UpperCamelCase ) is not None: return prompt_or_repo_id __UpperCAmelCase : str = cached_file( UpperCamelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(UpperCamelCase , "r" , encoding="utf-8" ) as f: return f.read()
77
1
"""simple docstring""" import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors A = logging.getLogger(__name__) class a__ ( __magic_name__ ): lowercase_ = "sequence-classification" def __init__( self : Optional[Any] , UpperCamelCase_ : int): """simple docstring""" if type(UpperCamelCase_) == dict: __UpperCAmelCase : Optional[int] = Namespace(**UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = glue_output_modes[hparams.task] __UpperCAmelCase : Optional[Any] = glue_tasks_num_labels[hparams.task] super().__init__(UpperCamelCase_ , UpperCamelCase_ , self.mode) def a_ ( self : Optional[Any] , **UpperCamelCase_ : Optional[int]): """simple docstring""" return self.model(**UpperCamelCase_) def a_ ( self : Optional[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Optional[int] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase : str = batch[2] if self.config.model_type in ["bert", "xlnet", "albert"] else None __UpperCAmelCase : Optional[Any] = self(**UpperCamelCase_) __UpperCAmelCase : int = outputs[0] __UpperCAmelCase : str = self.trainer.lr_schedulers[0]["scheduler"] __UpperCAmelCase : Optional[int] = {"loss": loss, "rate": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Any = self.hparams __UpperCAmelCase : List[Any] = processors[args.task]() __UpperCAmelCase : Dict = processor.get_labels() for mode in ["train", "dev"]: __UpperCAmelCase : str = self._feature_file(UpperCamelCase_) if os.path.exists(UpperCamelCase_) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , UpperCamelCase_) else: logger.info("Creating features from dataset file at %s" , args.data_dir) __UpperCAmelCase : Optional[int] = ( processor.get_dev_examples(args.data_dir) if mode == "dev" else processor.get_train_examples(args.data_dir) ) __UpperCAmelCase : List[Any] = convert_examples_to_features( UpperCamelCase_ , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("Saving features into cached file %s" , UpperCamelCase_) torch.save(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : str , UpperCamelCase_ : int , UpperCamelCase_ : bool = False): """simple docstring""" __UpperCAmelCase : Dict = "dev" if mode == "test" else mode __UpperCAmelCase : Dict = self._feature_file(UpperCamelCase_) logger.info("Loading features from cached file %s" , UpperCamelCase_) __UpperCAmelCase : Dict = torch.load(UpperCamelCase_) __UpperCAmelCase : List[str] = torch.tensor([f.input_ids for f in features] , dtype=torch.long) __UpperCAmelCase : int = torch.tensor([f.attention_mask for f in features] , dtype=torch.long) __UpperCAmelCase : Optional[Any] = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase : int = torch.tensor([f.label for f in features] , dtype=torch.long) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase : Optional[Any] = torch.tensor([f.label for f in features] , dtype=torch.float) return DataLoader( TensorDataset(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) , batch_size=UpperCamelCase_ , shuffle=UpperCamelCase_ , ) def a_ ( self : List[Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[str] = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase : List[Any] = batch[2] if self.config.model_type in ["bert", "xlnet", "albert"] else None __UpperCAmelCase : Any = self(**UpperCamelCase_) __UpperCAmelCase , __UpperCAmelCase : Tuple = outputs[:2] __UpperCAmelCase : List[Any] = logits.detach().cpu().numpy() __UpperCAmelCase : List[Any] = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def a_ ( self : List[Any] , UpperCamelCase_ : List[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.stack([x["val_loss"] for x in outputs]).mean().detach().cpu().item() __UpperCAmelCase : int = np.concatenate([x["pred"] for x in outputs] , axis=0) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase : str = np.argmax(UpperCamelCase_ , axis=1) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase : Union[str, Any] = np.squeeze(UpperCamelCase_) __UpperCAmelCase : List[str] = np.concatenate([x["target"] for x in outputs] , axis=0) __UpperCAmelCase : Optional[Any] = [[] for _ in range(out_label_ids.shape[0])] __UpperCAmelCase : Any = [[] for _ in range(out_label_ids.shape[0])] __UpperCAmelCase : Union[str, Any] = {**{"val_loss": val_loss_mean}, **compute_metrics(self.hparams.task , UpperCamelCase_ , UpperCamelCase_)} __UpperCAmelCase : Optional[int] = dict(results.items()) __UpperCAmelCase : str = results return ret, preds_list, out_label_list def a_ ( self : Union[str, Any] , UpperCamelCase_ : list): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = self._eval_end(UpperCamelCase_) __UpperCAmelCase : List[Any] = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def a_ ( self : List[str] , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = self._eval_end(UpperCamelCase_) __UpperCAmelCase : Any = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def a_ ( UpperCamelCase_ : Any , UpperCamelCase_ : List[Any]): """simple docstring""" BaseTransformer.add_model_specific_args(UpperCamelCase_ , UpperCamelCase_) parser.add_argument( "--max_seq_length" , default=128 , type=UpperCamelCase_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--task" , default="" , type=UpperCamelCase_ , required=UpperCamelCase_ , help="The GLUE task to run" , ) parser.add_argument( "--gpus" , default=0 , type=UpperCamelCase_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets") return parser def _UpperCamelCase ( ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : str = argparse.ArgumentParser() add_generic_args(UpperCamelCase , os.getcwd() ) __UpperCAmelCase : int = GLUETransformer.add_model_specific_args(UpperCamelCase , os.getcwd() ) __UpperCAmelCase : List[str] = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: __UpperCAmelCase : Union[str, Any] = os.path.join( "./results" , f"{args.task}_{time.strftime('%Y%m%d_%H%M%S' )}" , ) os.makedirs(args.output_dir ) __UpperCAmelCase : Tuple = GLUETransformer(UpperCamelCase ) __UpperCAmelCase : int = generic_train(UpperCamelCase , UpperCamelCase ) # Optionally, predict on dev set and write to output_dir if args.do_predict: __UpperCAmelCase : Union[str, Any] = sorted(glob.glob(os.path.join(args.output_dir , "checkpoint-epoch=*.ckpt" ) , recursive=UpperCamelCase ) ) __UpperCAmelCase : Tuple = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(UpperCamelCase ) if __name__ == "__main__": main()
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available A = { """configuration_ernie""": ["""ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ErnieConfig""", """ErnieOnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST""", """ErnieForCausalLM""", """ErnieForMaskedLM""", """ErnieForMultipleChoice""", """ErnieForNextSentencePrediction""", """ErnieForPreTraining""", """ErnieForQuestionAnswering""", """ErnieForSequenceClassification""", """ErnieForTokenClassification""", """ErnieModel""", """ErniePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase = 10**12 ) -> int: """simple docstring""" __UpperCAmelCase : Tuple = 1 __UpperCAmelCase : List[Any] = 0 __UpperCAmelCase : str = 1 __UpperCAmelCase : Union[str, Any] = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(f'''{solution() = }''')
77
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class a__ ( nn.Module ): def __init__( self : Union[str, Any]): """simple docstring""" super().__init__() __UpperCAmelCase : Optional[int] = nn.Linear(3 , 4) __UpperCAmelCase : str = nn.BatchNormad(4) __UpperCAmelCase : int = nn.Linear(4 , 5) def a_ ( self : str , UpperCamelCase_ : List[str]): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCamelCase_))) class a__ ( unittest.TestCase ): def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Optional[Any] = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , model.state_dict()) __UpperCAmelCase : Union[str, Any] = os.path.join(UpperCamelCase_ , "index.json") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: __UpperCAmelCase : Optional[int] = os.path.join(UpperCamelCase_ , F"{key}.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on the fact weights are properly loaded def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : int = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: __UpperCAmelCase : List[Any] = torch.randn(2 , 3 , dtype=UpperCamelCase_) with TemporaryDirectory() as tmp_dir: __UpperCAmelCase : Tuple = offload_weight(UpperCamelCase_ , "weight" , UpperCamelCase_ , {}) __UpperCAmelCase : Dict = os.path.join(UpperCamelCase_ , "weight.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) self.assertDictEqual(UpperCamelCase_ , {"weight": {"shape": [2, 3], "dtype": str(UpperCamelCase_).split(".")[1]}}) __UpperCAmelCase : Optional[Any] = load_offloaded_weight(UpperCamelCase_ , index["weight"]) self.assertTrue(torch.equal(UpperCamelCase_ , UpperCamelCase_)) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : List[Any] = ModelForTest() __UpperCAmelCase : Optional[int] = model.state_dict() __UpperCAmelCase : List[str] = {k: v for k, v in state_dict.items() if "linear2" not in k} __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "linear2" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : List[str] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "weight" in k} __UpperCAmelCase : Optional[Any] = {k: v for k, v in state_dict.items() if "weight" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[Any] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) # Duplicates are removed __UpperCAmelCase : str = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Any = {"a.1": 0, "a.10": 1, "a.2": 2} __UpperCAmelCase : Union[str, Any] = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1": 0, "a.2": 2}) __UpperCAmelCase : int = {"a.1.a": 0, "a.10.a": 1, "a.2.a": 2} __UpperCAmelCase : int = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1.a": 0, "a.2.a": 2})
77
1
"""simple docstring""" import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Tuple: """simple docstring""" __UpperCAmelCase : Any = AutoConfig.from_pretrained(UpperCamelCase ) __UpperCAmelCase : int = FlaxAutoModelForSeqaSeqLM.from_config(config=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = checkpoints.load_tax_checkpoint(UpperCamelCase ) __UpperCAmelCase : List[str] = "wi_0" in tax_model["target"]["encoder"]["layers_0"]["mlp"] if config.model_type == "t5": __UpperCAmelCase : Tuple = "SelfAttention" if config.model_type == "longt5" and config.encoder_attention_type == "local": __UpperCAmelCase : Optional[Any] = "LocalSelfAttention" elif config.model_type == "longt5" and config.encoder_attention_type == "transient-global": __UpperCAmelCase : Any = "TransientGlobalSelfAttention" else: raise ValueError( "Given config is expected to have `model_type='t5'`, or `model_type='longt5` with `encoder_attention_type`" " attribute with a value from ['local', 'transient-global]." ) # Encoder for layer_index in range(config.num_layers ): __UpperCAmelCase : Optional[Any] = f"layers_{str(UpperCamelCase )}" # Self-Attention __UpperCAmelCase : Optional[int] = tax_model["target"]["encoder"][layer_name]["attention"]["key"]["kernel"] __UpperCAmelCase : Any = tax_model["target"]["encoder"][layer_name]["attention"]["out"]["kernel"] __UpperCAmelCase : List[str] = tax_model["target"]["encoder"][layer_name]["attention"]["query"]["kernel"] __UpperCAmelCase : Dict = tax_model["target"]["encoder"][layer_name]["attention"]["value"]["kernel"] # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": __UpperCAmelCase : int = tax_model["target"]["encoder"][layer_name]["attention"]["T5LayerNorm_0"]["scale"] # Layer Normalization __UpperCAmelCase : str = tax_model["target"]["encoder"][layer_name]["pre_attention_layer_norm"]["scale"] if split_mlp_wi: __UpperCAmelCase : int = tax_model["target"]["encoder"][layer_name]["mlp"]["wi_0"]["kernel"] __UpperCAmelCase : List[str] = tax_model["target"]["encoder"][layer_name]["mlp"]["wi_1"]["kernel"] else: __UpperCAmelCase : Union[str, Any] = tax_model["target"]["encoder"][layer_name]["mlp"]["wi"]["kernel"] __UpperCAmelCase : Optional[int] = tax_model["target"]["encoder"][layer_name]["mlp"]["wo"]["kernel"] # Layer Normalization __UpperCAmelCase : str = tax_model["target"]["encoder"][layer_name]["pre_mlp_layer_norm"]["scale"] # Assigning __UpperCAmelCase : Any = flax_model.params["encoder"]["block"][str(UpperCamelCase )]["layer"] __UpperCAmelCase : int = tax_attention_key __UpperCAmelCase : Union[str, Any] = tax_attention_out __UpperCAmelCase : Optional[int] = tax_attention_query __UpperCAmelCase : Optional[Any] = tax_attention_value __UpperCAmelCase : int = tax_attention_layer_norm # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": __UpperCAmelCase : Any = tax_global_layer_norm if split_mlp_wi: __UpperCAmelCase : Any = tax_mlp_wi_a __UpperCAmelCase : Any = tax_mlp_wi_a else: __UpperCAmelCase : int = tax_mlp_wi __UpperCAmelCase : str = tax_mlp_wo __UpperCAmelCase : Optional[Any] = tax_mlp_layer_norm __UpperCAmelCase : Union[str, Any] = flax_model_encoder_layer_block # Only for layer 0: __UpperCAmelCase : Optional[Any] = tax_model["target"]["encoder"]["relpos_bias"]["rel_embedding"].T __UpperCAmelCase : int = tax_encoder_rel_embedding # Side/global relative position_bias + layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": __UpperCAmelCase : str = tax_model["target"]["encoder"]["side_relpos_bias"]["rel_embedding"].T __UpperCAmelCase : Any = tax_encoder_global_rel_embedding # Assigning __UpperCAmelCase : Any = tax_model["target"]["encoder"]["encoder_norm"]["scale"] __UpperCAmelCase : Dict = tax_encoder_norm # Decoder for layer_index in range(config.num_layers ): __UpperCAmelCase : List[str] = f"layers_{str(UpperCamelCase )}" # Self-Attention __UpperCAmelCase : str = tax_model["target"]["decoder"][layer_name]["self_attention"]["key"]["kernel"] __UpperCAmelCase : Any = tax_model["target"]["decoder"][layer_name]["self_attention"]["out"]["kernel"] __UpperCAmelCase : Dict = tax_model["target"]["decoder"][layer_name]["self_attention"]["query"]["kernel"] __UpperCAmelCase : Optional[Any] = tax_model["target"]["decoder"][layer_name]["self_attention"]["value"]["kernel"] # Layer Normalization __UpperCAmelCase : Optional[int] = tax_model["target"]["decoder"][layer_name]["pre_self_attention_layer_norm"][ "scale" ] # Encoder-Decoder-Attention __UpperCAmelCase : Any = tax_model["target"]["decoder"][layer_name]["encoder_decoder_attention"] __UpperCAmelCase : List[str] = tax_enc_dec_attention_module["key"]["kernel"] __UpperCAmelCase : List[str] = tax_enc_dec_attention_module["out"]["kernel"] __UpperCAmelCase : str = tax_enc_dec_attention_module["query"]["kernel"] __UpperCAmelCase : str = tax_enc_dec_attention_module["value"]["kernel"] # Layer Normalization __UpperCAmelCase : Optional[int] = tax_model["target"]["decoder"][layer_name]["pre_cross_attention_layer_norm"]["scale"] # MLP if split_mlp_wi: __UpperCAmelCase : Tuple = tax_model["target"]["decoder"][layer_name]["mlp"]["wi_0"]["kernel"] __UpperCAmelCase : str = tax_model["target"]["decoder"][layer_name]["mlp"]["wi_1"]["kernel"] else: __UpperCAmelCase : str = tax_model["target"]["decoder"][layer_name]["mlp"]["wi"]["kernel"] __UpperCAmelCase : Dict = tax_model["target"]["decoder"][layer_name]["mlp"]["wo"]["kernel"] # Layer Normalization __UpperCAmelCase : Tuple = tax_model["target"]["decoder"][layer_name]["pre_mlp_layer_norm"]["scale"] # Assigning __UpperCAmelCase : Union[str, Any] = flax_model.params["decoder"]["block"][str(UpperCamelCase )]["layer"] __UpperCAmelCase : Tuple = tax_attention_key __UpperCAmelCase : List[Any] = tax_attention_out __UpperCAmelCase : List[Any] = tax_attention_query __UpperCAmelCase : Optional[Any] = tax_attention_value __UpperCAmelCase : Union[str, Any] = tax_pre_attention_layer_norm __UpperCAmelCase : str = tax_enc_dec_attention_key __UpperCAmelCase : Any = tax_enc_dec_attention_out __UpperCAmelCase : str = tax_enc_dec_attention_query __UpperCAmelCase : str = tax_enc_dec_attention_value __UpperCAmelCase : List[Any] = tax_cross_layer_norm if split_mlp_wi: __UpperCAmelCase : List[str] = tax_mlp_wi_a __UpperCAmelCase : Optional[Any] = tax_mlp_wi_a else: __UpperCAmelCase : int = tax_mlp_wi __UpperCAmelCase : str = tax_mlp_wo __UpperCAmelCase : Any = txa_mlp_layer_norm __UpperCAmelCase : Optional[int] = flax_model_decoder_layer_block # Decoder Normalization __UpperCAmelCase : Union[str, Any] = tax_model["target"]["decoder"]["decoder_norm"]["scale"] __UpperCAmelCase : Dict = txa_decoder_norm # Only for layer 0: __UpperCAmelCase : Optional[int] = tax_model["target"]["decoder"]["relpos_bias"]["rel_embedding"].T __UpperCAmelCase : List[str] = tax_decoder_rel_embedding # Token Embeddings __UpperCAmelCase : Dict = tax_model["target"]["token_embedder"]["embedding"] __UpperCAmelCase : Optional[Any] = txa_token_embeddings # LM Head (only in v1.1 and LongT5 checkpoints) if "logits_dense" in tax_model["target"]["decoder"]: __UpperCAmelCase : List[str] = tax_model["target"]["decoder"]["logits_dense"]["kernel"] flax_model.save_pretrained(UpperCamelCase ) print("T5X Model was sucessfully converted!" ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path the T5X checkpoint.""" ) parser.add_argument("""--config_name""", default=None, type=str, required=True, help="""Config name of LongT5/T5 model.""") parser.add_argument( """--flax_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output FLAX model.""" ) A = parser.parse_args() convert_tax_checkpoint_to_flax(args.tax_checkpoint_path, args.config_name, args.flax_dump_folder_path)
77
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Dict = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __UpperCAmelCase : Union[str, Any] = n - k # Calculate C(n,k) for i in range(UpperCamelCase ): result *= n - i result //= i + 1 return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , UpperCamelCase ) // (node_count + 1) def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if n < 0: raise ValueError("factorial() not defined for negative values" ) __UpperCAmelCase : Optional[Any] = 1 for i in range(1 , n + 1 ): result *= i return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return catalan_number(UpperCamelCase ) * factorial(UpperCamelCase ) if __name__ == "__main__": A = int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Dict = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __UpperCAmelCase : Union[str, Any] = n - k # Calculate C(n,k) for i in range(UpperCamelCase ): result *= n - i result //= i + 1 return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , UpperCamelCase ) // (node_count + 1) def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if n < 0: raise ValueError("factorial() not defined for negative values" ) __UpperCAmelCase : Optional[Any] = 1 for i in range(1 , n + 1 ): result *= i return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return catalan_number(UpperCamelCase ) * factorial(UpperCamelCase ) if __name__ == "__main__": A = int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) A = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { """uw-madison/mra-base-512-4""": """https://huggingface.co/uw-madison/mra-base-512-4/resolve/main/config.json""", } class a__ ( __magic_name__ ): lowercase_ = "mra" def __init__( self : Any , UpperCamelCase_ : List[Any]=50265 , UpperCamelCase_ : List[Any]=768 , UpperCamelCase_ : Dict=12 , UpperCamelCase_ : int=12 , UpperCamelCase_ : str=3072 , UpperCamelCase_ : Optional[int]="gelu" , UpperCamelCase_ : Union[str, Any]=0.1 , UpperCamelCase_ : List[Any]=0.1 , UpperCamelCase_ : Dict=512 , UpperCamelCase_ : Optional[int]=1 , UpperCamelCase_ : Optional[Any]=0.02 , UpperCamelCase_ : List[str]=1e-5 , UpperCamelCase_ : str="absolute" , UpperCamelCase_ : Tuple=4 , UpperCamelCase_ : Any="full" , UpperCamelCase_ : List[Any]=0 , UpperCamelCase_ : List[Any]=0 , UpperCamelCase_ : Tuple=1 , UpperCamelCase_ : Dict=0 , UpperCamelCase_ : int=2 , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" super().__init__(pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : str = vocab_size __UpperCAmelCase : int = max_position_embeddings __UpperCAmelCase : Dict = hidden_size __UpperCAmelCase : List[str] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : Tuple = hidden_act __UpperCAmelCase : Optional[Any] = hidden_dropout_prob __UpperCAmelCase : int = attention_probs_dropout_prob __UpperCAmelCase : Union[str, Any] = initializer_range __UpperCAmelCase : List[Any] = type_vocab_size __UpperCAmelCase : Dict = layer_norm_eps __UpperCAmelCase : Optional[Any] = position_embedding_type __UpperCAmelCase : Any = block_per_row __UpperCAmelCase : Optional[int] = approx_mode __UpperCAmelCase : int = initial_prior_first_n_blocks __UpperCAmelCase : Tuple = initial_prior_diagonal_n_blocks
77
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCamelCase_ : int , UpperCamelCase_ : Any=13 , UpperCamelCase_ : Optional[int]=3 , UpperCamelCase_ : int=224 , UpperCamelCase_ : int=30 , UpperCamelCase_ : str=400 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Optional[int]=[0.5, 0.5, 0.5] , UpperCamelCase_ : Optional[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __UpperCAmelCase : Tuple = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : List[Any] = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : List[Any] = image_size __UpperCAmelCase : str = min_resolution __UpperCAmelCase : Tuple = max_resolution __UpperCAmelCase : Optional[Any] = do_resize __UpperCAmelCase : Any = size __UpperCAmelCase : Any = do_normalize __UpperCAmelCase : Any = image_mean __UpperCAmelCase : Optional[Any] = image_std def a_ ( self : str): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ViTImageProcessor if is_vision_available() else None def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Optional[Any] = EfficientFormerImageProcessorTester(self) @property def a_ ( self : Union[str, Any]): """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCamelCase_ , "image_std")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) def a_ ( self : Dict): """simple docstring""" pass def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : str = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Optional[int] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : Tuple = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Any = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available A = {"""tokenization_herbert""": ["""HerbertTokenizer"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""HerbertTokenizerFast"""] if TYPE_CHECKING: from .tokenization_herbert import HerbertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_herbert_fast import HerbertTokenizerFast else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" from collections import namedtuple A = namedtuple("""from_to""", """from_ to""") A = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1_000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": from_to(0.76455, 1.30795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.000236588, 4226.75), } def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'from_type' value: {from_type!r} Supported values are:\n" + ", ".join(UpperCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n" + ", ".join(UpperCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
77
1
"""simple docstring""" import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConformerConfig, WavaVecaConformerForCTC, WavaVecaConformerForPreTraining, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() A = logging.get_logger(__name__) A = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.linear_k""": """encoder.layers.*.self_attn.linear_k""", """self_attn.linear_v""": """encoder.layers.*.self_attn.linear_v""", """self_attn.linear_q""": """encoder.layers.*.self_attn.linear_q""", """self_attn.pos_bias_u""": """encoder.layers.*.self_attn.pos_bias_u""", """self_attn.pos_bias_v""": """encoder.layers.*.self_attn.pos_bias_v""", """self_attn.linear_out""": """encoder.layers.*.self_attn.linear_out""", """self_attn.linear_pos""": """encoder.layers.*.self_attn.linear_pos""", """self_attn.rotary_emb""": """encoder.embed_positions""", """self_attn_layer_norm""": """encoder.layers.*.self_attn_layer_norm""", """conv_module.pointwise_conv1""": """encoder.layers.*.conv_module.pointwise_conv1""", """conv_module.pointwise_conv2""": """encoder.layers.*.conv_module.pointwise_conv2""", """conv_module.depthwise_conv""": """encoder.layers.*.conv_module.depthwise_conv""", """conv_module.batch_norm""": """encoder.layers.*.conv_module.batch_norm""", """conv_module.layer_norm""": """encoder.layers.*.conv_module.layer_norm""", """ffn1.w_1""": """encoder.layers.*.ffn1.intermediate_dense""", """ffn1.w_2""": """encoder.layers.*.ffn1.output_dense""", """ffn1.layer_norm""": """encoder.layers.*.ffn1_layer_norm""", """ffn2.w_1""": """encoder.layers.*.ffn2.intermediate_dense""", """ffn2.w_2""": """encoder.layers.*.ffn2.output_dense""", """ffn2.layer_norm""": """encoder.layers.*.ffn2_layer_norm""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """lm_head""", """mask_emb""": """masked_spec_embed""", } A = [ """lm_head""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" for attribute in key.split("." ): __UpperCAmelCase : List[str] = getattr(UpperCamelCase , UpperCamelCase ) if weight_type is not None: __UpperCAmelCase : List[str] = getattr(UpperCamelCase , UpperCamelCase ).shape else: __UpperCAmelCase : Optional[int] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be" f" {value.shape} for {full_name}" ) if weight_type == "weight": __UpperCAmelCase : Union[str, Any] = value elif weight_type == "weight_g": __UpperCAmelCase : List[Any] = value elif weight_type == "weight_v": __UpperCAmelCase : Optional[int] = value elif weight_type == "bias": __UpperCAmelCase : int = value elif weight_type == "running_mean": __UpperCAmelCase : int = value elif weight_type == "running_var": __UpperCAmelCase : Optional[int] = value elif weight_type == "num_batches_tracked": __UpperCAmelCase : Any = value elif weight_type == "inv_freq": __UpperCAmelCase : List[str] = value else: __UpperCAmelCase : Union[str, Any] = value logger.info(f"{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}." ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[Any]: """simple docstring""" __UpperCAmelCase : Union[str, Any] = [] __UpperCAmelCase : Optional[int] = fairseq_model.state_dict() __UpperCAmelCase : str = hf_model.wavaveca_conformer.feature_extractor for name, value in fairseq_dict.items(): __UpperCAmelCase : Dict = False if "conv_layers" in name: load_conv_layer( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , hf_model.config.feat_extract_norm == "group" , ) __UpperCAmelCase : Optional[Any] = True else: for key, mapped_key in MAPPING.items(): __UpperCAmelCase : List[Any] = "wav2vec2_conformer." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: __UpperCAmelCase : str = True if "*" in mapped_key: __UpperCAmelCase : Tuple = name.split(UpperCamelCase )[0].split("." )[-2] __UpperCAmelCase : Any = mapped_key.replace("*" , UpperCamelCase ) if "pos_bias_u" in name: __UpperCAmelCase : Dict = None elif "pos_bias_v" in name: __UpperCAmelCase : Optional[Any] = None elif "weight_g" in name: __UpperCAmelCase : Union[str, Any] = "weight_g" elif "weight_v" in name: __UpperCAmelCase : Dict = "weight_v" elif "bias" in name: __UpperCAmelCase : Tuple = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj __UpperCAmelCase : Tuple = "weight" elif "running_mean" in name: __UpperCAmelCase : int = "running_mean" elif "inv_freq" in name: __UpperCAmelCase : Tuple = "inv_freq" elif "running_var" in name: __UpperCAmelCase : Union[str, Any] = "running_var" elif "num_batches_tracked" in name: __UpperCAmelCase : Optional[int] = "num_batches_tracked" else: __UpperCAmelCase : List[str] = None set_recursively(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) continue if not is_used: unused_weights.append(UpperCamelCase ) logger.warning(f"Unused weights: {unused_weights}" ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : Optional[Any] = full_name.split("conv_layers." )[-1] __UpperCAmelCase : List[Any] = name.split("." ) __UpperCAmelCase : List[Any] = int(items[0] ) __UpperCAmelCase : List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found." ) __UpperCAmelCase : Union[str, Any] = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found." ) __UpperCAmelCase : Dict = value logger.info(f"Feat extract conv layer {layer_id} was initialized from {full_name}." ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found." ) __UpperCAmelCase : int = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"{full_name} has size {value.shape}, but" f" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found." ) __UpperCAmelCase : Any = value logger.info(f"Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}." ) else: unused_weights.append(UpperCamelCase ) @torch.no_grad() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase=None , UpperCamelCase=None , UpperCamelCase=True ) -> Optional[int]: """simple docstring""" if config_path is not None: __UpperCAmelCase : int = WavaVecaConformerConfig.from_pretrained(UpperCamelCase , hidden_act="swish" ) else: __UpperCAmelCase : List[Any] = WavaVecaConformerConfig() if "rope" in checkpoint_path: __UpperCAmelCase : str = "rotary" if is_finetuned: if dict_path: __UpperCAmelCase : int = Dictionary.load(UpperCamelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __UpperCAmelCase : Optional[Any] = target_dict.pad_index __UpperCAmelCase : Dict = target_dict.bos_index __UpperCAmelCase : Dict = target_dict.eos_index __UpperCAmelCase : List[Any] = len(target_dict.symbols ) __UpperCAmelCase : Union[str, Any] = os.path.join(UpperCamelCase , "vocab.json" ) if not os.path.isdir(UpperCamelCase ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(UpperCamelCase ) ) return os.makedirs(UpperCamelCase , exist_ok=UpperCamelCase ) __UpperCAmelCase : Dict = target_dict.indices # fairseq has the <pad> and <s> switched __UpperCAmelCase : Any = 0 __UpperCAmelCase : Tuple = 1 with open(UpperCamelCase , "w" , encoding="utf-8" ) as vocab_handle: json.dump(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Any = WavaVecaCTCTokenizer( UpperCamelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=UpperCamelCase , ) __UpperCAmelCase : List[str] = True if config.feat_extract_norm == "layer" else False __UpperCAmelCase : str = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=UpperCamelCase , return_attention_mask=UpperCamelCase , ) __UpperCAmelCase : List[Any] = WavaVecaProcessor(feature_extractor=UpperCamelCase , tokenizer=UpperCamelCase ) processor.save_pretrained(UpperCamelCase ) __UpperCAmelCase : Any = WavaVecaConformerForCTC(UpperCamelCase ) else: __UpperCAmelCase : int = WavaVecaConformerForPreTraining(UpperCamelCase ) if is_finetuned: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : str = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] )} ) else: __UpperCAmelCase : int = argparse.Namespace(task="audio_pretraining" ) __UpperCAmelCase : List[str] = fairseq.tasks.setup_task(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=UpperCamelCase ) __UpperCAmelCase : List[str] = model[0].eval() recursively_load_weights(UpperCamelCase , UpperCamelCase , not is_finetuned ) hf_wavavec.save_pretrained(UpperCamelCase ) if __name__ == "__main__": A = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) A = parser.parse_args() convert_wavaveca_conformer_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
77
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ShapEPipeline lowercase_ = ["prompt"] lowercase_ = ["prompt"] lowercase_ = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] lowercase_ = False @property def a_ ( self : Optional[int]): """simple docstring""" return 32 @property def a_ ( self : Any): """simple docstring""" return 32 @property def a_ ( self : int): """simple docstring""" return self.time_input_dim * 4 @property def a_ ( self : List[Any]): """simple docstring""" return 8 @property def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") return tokenizer @property def a_ ( self : List[str]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(UpperCamelCase_) @property def a_ ( self : Any): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Union[str, Any] = { "num_attention_heads": 2, "attention_head_dim": 16, "embedding_dim": self.time_input_dim, "num_embeddings": 32, "embedding_proj_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "num_layers": 1, "clip_embed_dim": self.time_input_dim * 2, "additional_embeddings": 0, "time_embed_act_fn": "gelu", "norm_in_type": "layer", "encoder_hid_proj_type": None, "added_emb_type": None, } __UpperCAmelCase : Dict = PriorTransformer(**UpperCamelCase_) return model @property def a_ ( self : Union[str, Any]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Tuple = { "param_shapes": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), "d_latent": self.time_input_dim, "d_hidden": self.renderer_dim, "n_output": 12, "background": ( 0.1, 0.1, 0.1, ), } __UpperCAmelCase : List[Any] = ShapERenderer(**UpperCamelCase_) return model def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.dummy_prior __UpperCAmelCase : str = self.dummy_text_encoder __UpperCAmelCase : int = self.dummy_tokenizer __UpperCAmelCase : int = self.dummy_renderer __UpperCAmelCase : Tuple = HeunDiscreteScheduler( beta_schedule="exp" , num_train_timesteps=1024 , prediction_type="sample" , use_karras_sigmas=UpperCamelCase_ , clip_sample=UpperCamelCase_ , clip_sample_range=1.0 , ) __UpperCAmelCase : str = { "prior": prior, "text_encoder": text_encoder, "tokenizer": tokenizer, "renderer": renderer, "scheduler": scheduler, } return components def a_ ( self : Optional[int] , UpperCamelCase_ : Dict , UpperCamelCase_ : Any=0): """simple docstring""" if str(UpperCamelCase_).startswith("mps"): __UpperCAmelCase : List[Any] = torch.manual_seed(UpperCamelCase_) else: __UpperCAmelCase : str = torch.Generator(device=UpperCamelCase_).manual_seed(UpperCamelCase_) __UpperCAmelCase : List[Any] = { "prompt": "horse", "generator": generator, "num_inference_steps": 1, "frame_size": 32, "output_type": "np", } return inputs def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : str = "cpu" __UpperCAmelCase : Union[str, Any] = self.get_dummy_components() __UpperCAmelCase : Union[str, Any] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[Any] = pipe(**self.get_dummy_inputs(UpperCamelCase_)) __UpperCAmelCase : Union[str, Any] = output.images[0] __UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __UpperCAmelCase : Union[str, Any] = np.array( [ 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def a_ ( self : Tuple): """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2]) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch_device == "cpu" __UpperCAmelCase : Optional[Any] = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=UpperCamelCase_ , relax_max_difference=UpperCamelCase_ , ) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.get_dummy_components() __UpperCAmelCase : List[str] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : int = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[int] = 1 __UpperCAmelCase : Any = 2 __UpperCAmelCase : Optional[Any] = self.get_dummy_inputs(UpperCamelCase_) for key in inputs.keys(): if key in self.batch_params: __UpperCAmelCase : List[Any] = batch_size * [inputs[key]] __UpperCAmelCase : List[Any] = pipe(**UpperCamelCase_ , num_images_per_prompt=UpperCamelCase_)[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class a__ ( unittest.TestCase ): def a_ ( self : List[str]): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/test_shap_e_np_out.npy") __UpperCAmelCase : Optional[Any] = ShapEPipeline.from_pretrained("openai/shap-e") __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Dict = torch.Generator(device=UpperCamelCase_).manual_seed(0) __UpperCAmelCase : int = pipe( "a shark" , generator=UpperCamelCase_ , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type="np" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(UpperCamelCase_ , UpperCamelCase_)
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { """google/vivit-b-16x2-kinetics400""": ( """https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json""" ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class a__ ( __magic_name__ ): lowercase_ = "vivit" def __init__( self : Union[str, Any] , UpperCamelCase_ : Dict=224 , UpperCamelCase_ : Optional[Any]=32 , UpperCamelCase_ : List[Any]=[2, 16, 16] , UpperCamelCase_ : str=3 , UpperCamelCase_ : Tuple=768 , UpperCamelCase_ : Tuple=12 , UpperCamelCase_ : Any=12 , UpperCamelCase_ : Any=3072 , UpperCamelCase_ : int="gelu_fast" , UpperCamelCase_ : Any=0.0 , UpperCamelCase_ : Tuple=0.0 , UpperCamelCase_ : Optional[int]=0.02 , UpperCamelCase_ : Tuple=1e-06 , UpperCamelCase_ : Optional[int]=True , **UpperCamelCase_ : int , ): """simple docstring""" __UpperCAmelCase : List[Any] = hidden_size __UpperCAmelCase : Optional[Any] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : Tuple = intermediate_size __UpperCAmelCase : int = hidden_act __UpperCAmelCase : Tuple = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : Optional[Any] = initializer_range __UpperCAmelCase : List[Any] = layer_norm_eps __UpperCAmelCase : Dict = image_size __UpperCAmelCase : Optional[Any] = num_frames __UpperCAmelCase : List[Any] = tubelet_size __UpperCAmelCase : Optional[Any] = num_channels __UpperCAmelCase : Optional[int] = qkv_bias super().__init__(**UpperCamelCase_)
77
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_features", "is_longer"] def __init__( self : List[str] , UpperCamelCase_ : Dict=64 , UpperCamelCase_ : Tuple=48000 , UpperCamelCase_ : List[Any]=480 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=1024 , UpperCamelCase_ : List[str]=0.0 , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : float = 0 , UpperCamelCase_ : float = 14000 , UpperCamelCase_ : int = None , UpperCamelCase_ : str = "fusion" , UpperCamelCase_ : str = "repeatpad" , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" super().__init__( feature_size=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , padding_value=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = top_db __UpperCAmelCase : Optional[Any] = truncation __UpperCAmelCase : str = padding __UpperCAmelCase : int = fft_window_size __UpperCAmelCase : str = (fft_window_size >> 1) + 1 __UpperCAmelCase : List[Any] = hop_length __UpperCAmelCase : Optional[Any] = max_length_s __UpperCAmelCase : Tuple = max_length_s * sampling_rate __UpperCAmelCase : str = sampling_rate __UpperCAmelCase : int = frequency_min __UpperCAmelCase : Optional[Any] = frequency_max __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm=UpperCamelCase_ , mel_scale="htk" , ) __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm="slaney" , mel_scale="slaney" , ) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Dict = copy.deepcopy(self.__dict__) __UpperCAmelCase : str = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def a_ ( self : int , UpperCamelCase_ : np.array , UpperCamelCase_ : Optional[np.array] = None): """simple docstring""" __UpperCAmelCase : List[Any] = spectrogram( UpperCamelCase_ , window_function(self.fft_window_size , "hann") , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=UpperCamelCase_ , log_mel="dB" , ) return log_mel_spectrogram.T def a_ ( self : List[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = np.array_split(list(range(0 , total_frames - chunk_frames + 1)) , 3) if len(ranges[1]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : str = [0] if len(ranges[2]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : Dict = [0] # randomly choose index for each part __UpperCAmelCase : Dict = np.random.choice(ranges[0]) __UpperCAmelCase : List[str] = np.random.choice(ranges[1]) __UpperCAmelCase : List[Any] = np.random.choice(ranges[2]) __UpperCAmelCase : List[Any] = mel[idx_front : idx_front + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_back : idx_back + chunk_frames, :] __UpperCAmelCase : Tuple = torch.tensor(mel[None, None, :]) __UpperCAmelCase : Union[str, Any] = torch.nn.functional.interpolate( UpperCamelCase_ , size=[chunk_frames, 64] , mode="bilinear" , align_corners=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = mel_shrink[0][0].numpy() __UpperCAmelCase : Optional[int] = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0) return mel_fusion def a_ ( self : Optional[Any] , UpperCamelCase_ : np.array , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[Any]): """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": __UpperCAmelCase : List[str] = True # random crop to max_length (for compatibility) -> this should be handled by self.pad __UpperCAmelCase : List[Any] = len(UpperCamelCase_) - max_length __UpperCAmelCase : int = np.random.randint(0 , overflow + 1) __UpperCAmelCase : Union[str, Any] = waveform[idx : idx + max_length] __UpperCAmelCase : Union[str, Any] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] elif truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : Dict = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed __UpperCAmelCase : Tuple = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. __UpperCAmelCase : List[str] = np.stack([mel, mel, mel, mel] , axis=0) __UpperCAmelCase : Any = False else: __UpperCAmelCase : List[str] = self._random_mel_fusion(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented") else: __UpperCAmelCase : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": __UpperCAmelCase : Tuple = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : List[str] = np.stack(np.tile(UpperCamelCase_ , n_repeat + 1))[:max_length] if padding == "repeatpad": __UpperCAmelCase : Union[str, Any] = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : Optional[Any] = np.stack(np.tile(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : int = np.pad(UpperCamelCase_ , (0, max_length - waveform.shape[0]) , mode="constant" , constant_values=0) if truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0) else: __UpperCAmelCase : Optional[int] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] return input_mel, longer def __call__( self : Dict , UpperCamelCase_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase_ : str = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[Union[str, TensorType]] = None , **UpperCamelCase_ : Any , ): """simple docstring""" __UpperCAmelCase : int = truncation if truncation is not None else self.truncation __UpperCAmelCase : Optional[Any] = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}.") else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug.") __UpperCAmelCase : List[str] = isinstance(UpperCamelCase_ , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}") __UpperCAmelCase : str = is_batched_numpy or ( isinstance(UpperCamelCase_ , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: __UpperCAmelCase : Dict = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase_ , np.ndarray): __UpperCAmelCase : Tuple = np.asarray(UpperCamelCase_ , dtype=np.floataa) elif isinstance(UpperCamelCase_ , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): __UpperCAmelCase : Optional[int] = raw_speech.astype(np.floataa) # always return batch if not is_batched: __UpperCAmelCase : int = [np.asarray(UpperCamelCase_)] # convert to mel spectrogram, truncate and pad if needed. __UpperCAmelCase : Optional[int] = [ self._get_input_mel(UpperCamelCase_ , max_length if max_length else self.nb_max_samples , UpperCamelCase_ , UpperCamelCase_) for waveform in raw_speech ] __UpperCAmelCase : Tuple = [] __UpperCAmelCase : List[Any] = [] for mel, longer in padded_inputs: input_mel.append(UpperCamelCase_) is_longer.append(UpperCamelCase_) if truncation == "fusion" and sum(UpperCamelCase_) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer __UpperCAmelCase : Any = np.random.randint(0 , len(UpperCamelCase_)) __UpperCAmelCase : Optional[int] = True if isinstance(input_mel[0] , UpperCamelCase_): __UpperCAmelCase : Tuple = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for feature in input_mel] # is_longer is a list of bool __UpperCAmelCase : List[str] = [[longer] for longer in is_longer] __UpperCAmelCase : Optional[int] = {"input_features": input_mel, "is_longer": is_longer} __UpperCAmelCase : Optional[int] = BatchFeature(UpperCamelCase_) if return_tensors is not None: __UpperCAmelCase : Any = input_features.convert_to_tensors(UpperCamelCase_) return input_features
77
1
"""simple docstring""" from math import factorial class a__ : def __init__( self : Optional[int] , UpperCamelCase_ : str , UpperCamelCase_ : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = real if isinstance(UpperCamelCase_ , UpperCamelCase_): __UpperCAmelCase : Any = [1] * rank else: __UpperCAmelCase : List[str] = rank def __repr__( self : Dict): """simple docstring""" return ( F"{self.real}+" F"{'+'.join(str(UpperCamelCase_)+'E'+str(n+1)for n,dual in enumerate(self.duals))}" ) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.duals.copy() while cur[-1] == 0: cur.pop(-1) return Dual(self.real , UpperCamelCase_) def __add__( self : Optional[Any] , UpperCamelCase_ : Tuple): """simple docstring""" if not isinstance(UpperCamelCase_ , UpperCamelCase_): return Dual(self.real + other , self.duals) __UpperCAmelCase : List[str] = self.duals.copy() __UpperCAmelCase : Dict = other.duals.copy() if len(UpperCamelCase_) > len(UpperCamelCase_): o_dual.extend([1] * (len(UpperCamelCase_) - len(UpperCamelCase_))) elif len(UpperCamelCase_) < len(UpperCamelCase_): s_dual.extend([1] * (len(UpperCamelCase_) - len(UpperCamelCase_))) __UpperCAmelCase : int = [] for i in range(len(UpperCamelCase_)): new_duals.append(s_dual[i] + o_dual[i]) return Dual(self.real + other.real , UpperCamelCase_) lowercase_ = __add__ def __sub__( self : List[Any] , UpperCamelCase_ : int): """simple docstring""" return self + other * -1 def __mul__( self : List[str] , UpperCamelCase_ : Tuple): """simple docstring""" if not isinstance(UpperCamelCase_ , UpperCamelCase_): __UpperCAmelCase : str = [] for i in self.duals: new_duals.append(i * other) return Dual(self.real * other , UpperCamelCase_) __UpperCAmelCase : str = [0] * (len(self.duals) + len(other.duals) + 1) for i, item in enumerate(self.duals): for j, jtem in enumerate(other.duals): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals)): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals)): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real , UpperCamelCase_) lowercase_ = __mul__ def __truediv__( self : Union[str, Any] , UpperCamelCase_ : Any): """simple docstring""" if not isinstance(UpperCamelCase_ , UpperCamelCase_): __UpperCAmelCase : List[Any] = [] for i in self.duals: new_duals.append(i / other) return Dual(self.real / other , UpperCamelCase_) raise ValueError def __floordiv__( self : List[Any] , UpperCamelCase_ : Tuple): """simple docstring""" if not isinstance(UpperCamelCase_ , UpperCamelCase_): __UpperCAmelCase : Optional[int] = [] for i in self.duals: new_duals.append(i // other) return Dual(self.real // other , UpperCamelCase_) raise ValueError def __pow__( self : Dict , UpperCamelCase_ : Any): """simple docstring""" if n < 0 or isinstance(UpperCamelCase_ , UpperCamelCase_): raise ValueError("power must be a positive integer") if n == 0: return 1 if n == 1: return self __UpperCAmelCase : str = self for _ in range(n - 1): x *= self return x def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" if not callable(UpperCamelCase ): raise ValueError("differentiate() requires a function as input for func" ) if not isinstance(UpperCamelCase , (float, int) ): raise ValueError("differentiate() requires a float as input for position" ) if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("differentiate() requires an int as input for order" ) __UpperCAmelCase : Optional[Any] = Dual(UpperCamelCase , 1 ) __UpperCAmelCase : Any = func(UpperCamelCase ) if order == 0: return result.real return result.duals[order - 1] * factorial(UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() def _UpperCamelCase ( UpperCamelCase ) -> List[Any]: """simple docstring""" return y**2 * y**4 print(differentiate(f, 9, 2))
77
"""simple docstring""" import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , UpperCamelCase_ : List[Any]="</s>" , UpperCamelCase_ : Tuple="<unk>" , UpperCamelCase_ : List[str]="<pad>" , UpperCamelCase_ : Union[str, Any]=125 , UpperCamelCase_ : Dict=None , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: __UpperCAmelCase : int = [F"<extra_id_{i}>" for i in range(UpperCamelCase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens __UpperCAmelCase : Dict = len(set(filter(lambda UpperCamelCase_: bool("extra_id" in str(UpperCamelCase_)) , UpperCamelCase_))) if extra_tokens != extra_ids: raise ValueError( F"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" " provided to ByT5Tokenizer. In this case the additional_special_tokens must include the" " extra_ids tokens") __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else pad_token __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else eos_token __UpperCAmelCase : Optional[int] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else unk_token super().__init__( eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , extra_ids=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : List[str] = extra_ids __UpperCAmelCase : int = 2**8 # utf is 8 bits # define special tokens dict __UpperCAmelCase : Dict[int, str] = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } __UpperCAmelCase : Any = len(self.special_tokens_encoder) __UpperCAmelCase : List[Any] = len(UpperCamelCase_) for i, token in enumerate(UpperCamelCase_): __UpperCAmelCase : Union[str, Any] = self.vocab_size + i - n __UpperCAmelCase : Dict[str, int] = {v: k for k, v in self.special_tokens_encoder.items()} @property def a_ ( self : List[Any]): """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def a_ ( self : List[str] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(UpperCamelCase_)) + [1] return ([0] * len(UpperCamelCase_)) + [1] + ([0] * len(UpperCamelCase_)) + [1] def a_ ( self : Optional[Any] , UpperCamelCase_ : List[int]): """simple docstring""" if len(UpperCamelCase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F"This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated" " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def a_ ( self : Dict , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Dict = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def a_ ( self : Optional[int] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Optional[Any] = self._add_eos_if_not_present(UpperCamelCase_) if token_ids_a is None: return token_ids_a else: __UpperCAmelCase : List[Any] = self._add_eos_if_not_present(UpperCamelCase_) return token_ids_a + token_ids_a def a_ ( self : List[str] , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : Any = [chr(UpperCamelCase_) for i in text.encode("utf-8")] return tokens def a_ ( self : Tuple , UpperCamelCase_ : List[Any]): """simple docstring""" if token in self.special_tokens_encoder: __UpperCAmelCase : Any = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: __UpperCAmelCase : int = self.added_tokens_encoder[token] elif len(UpperCamelCase_) != 1: __UpperCAmelCase : Optional[Any] = self.unk_token_id else: __UpperCAmelCase : Any = ord(UpperCamelCase_) + self._num_special_tokens return token_id def a_ ( self : Any , UpperCamelCase_ : List[str]): """simple docstring""" if index in self.special_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[index] else: __UpperCAmelCase : List[str] = chr(index - self._num_special_tokens) return token def a_ ( self : Dict , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : str = b"" for token in tokens: if token in self.special_tokens_decoder: __UpperCAmelCase : Tuple = self.special_tokens_decoder[token].encode("utf-8") elif token in self.added_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[token].encode("utf-8") elif token in self.special_tokens_encoder: __UpperCAmelCase : Optional[int] = token.encode("utf-8") elif token in self.added_tokens_encoder: __UpperCAmelCase : Optional[Any] = token.encode("utf-8") else: __UpperCAmelCase : Any = bytes([ord(UpperCamelCase_)]) bstring += tok_string __UpperCAmelCase : List[Any] = bstring.decode("utf-8" , errors="ignore") return string def a_ ( self : Optional[int] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" return ()
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : int = [] __UpperCAmelCase : Optional[int] = [] __UpperCAmelCase : Union[str, Any] = { "^": 3, "*": 2, "/": 2, "%": 2, "+": 1, "-": 1, } # Priority of each operator __UpperCAmelCase : List[Any] = len(UpperCamelCase ) if (len(UpperCamelCase ) > 7) else 7 # Print table header for output print( "Symbol".center(8 ) , "Stack".center(UpperCamelCase ) , "Postfix".center(UpperCamelCase ) , sep=" | " , ) print("-" * (print_width * 3 + 7) ) for x in infix: if x.isalpha() or x.isdigit(): post_fix.append(UpperCamelCase ) # if x is Alphabet / Digit, add it to Postfix elif x == "(": stack.append(UpperCamelCase ) # if x is "(" push to Stack elif x == ")": # if x is ")" pop stack until "(" is encountered while stack[-1] != "(": post_fix.append(stack.pop() ) # Pop stack & add the content to Postfix stack.pop() else: if len(UpperCamelCase ) == 0: stack.append(UpperCamelCase ) # If stack is empty, push x to stack else: # while priority of x is not > priority of element in the stack while len(UpperCamelCase ) > 0 and priority[x] <= priority[stack[-1]]: post_fix.append(stack.pop() ) # pop stack & add to Postfix stack.append(UpperCamelCase ) # push x to stack print( x.center(8 ) , ("".join(UpperCamelCase )).ljust(UpperCamelCase ) , ("".join(UpperCamelCase )).ljust(UpperCamelCase ) , sep=" | " , ) # Output in tabular format while len(UpperCamelCase ) > 0: # while stack is not empty post_fix.append(stack.pop() ) # pop stack & add to Postfix print( " ".center(8 ) , ("".join(UpperCamelCase )).ljust(UpperCamelCase ) , ("".join(UpperCamelCase )).ljust(UpperCamelCase ) , sep=" | " , ) # Output in tabular format return "".join(UpperCamelCase ) # return Postfix as str def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" __UpperCAmelCase : Any = list(infix[::-1] ) # reverse the infix equation for i in range(len(UpperCamelCase ) ): if infix[i] == "(": __UpperCAmelCase : Optional[int] = ")" # change "(" to ")" elif infix[i] == ")": __UpperCAmelCase : int = "(" # change ")" to "(" return (infix_2_postfix("".join(UpperCamelCase ) ))[ ::-1 ] # call infix_2_postfix on Infix, return reverse of Postfix if __name__ == "__main__": A = input("""\nEnter an Infix Equation = """) # Input an Infix equation A = """""".join(Infix.split()) # Remove spaces from the input print("""\n\t""", Infix, """(Infix) -> """, infix_2_prefix(Infix), """(Prefix)""")
77
"""simple docstring""" import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a__ ( unittest.TestCase ): def __init__( self : List[Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple=3 , UpperCamelCase_ : Optional[int]=32 , UpperCamelCase_ : Dict=3 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=[10, 20, 30, 40] , UpperCamelCase_ : Tuple=[1, 1, 2, 1] , UpperCamelCase_ : str=True , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Dict="relu" , UpperCamelCase_ : str=3 , UpperCamelCase_ : int=None , ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : List[str] = image_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : Union[str, Any] = embeddings_size __UpperCAmelCase : Dict = hidden_sizes __UpperCAmelCase : Dict = depths __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_labels __UpperCAmelCase : Optional[int] = hidden_act __UpperCAmelCase : str = num_labels __UpperCAmelCase : Optional[int] = scope __UpperCAmelCase : Dict = len(UpperCamelCase_) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __UpperCAmelCase : Dict = self.get_config() return config, pixel_values def a_ ( self : Dict): """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def a_ ( self : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[str] = FlaxRegNetModel(config=UpperCamelCase_) __UpperCAmelCase : Dict = model(UpperCamelCase_) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a_ ( self : Any , UpperCamelCase_ : Dict , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : List[Any] = self.num_labels __UpperCAmelCase : Tuple = FlaxRegNetForImageClassification(config=UpperCamelCase_) __UpperCAmelCase : str = model(UpperCamelCase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Any = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase : Tuple = config_and_inputs __UpperCAmelCase : Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () lowercase_ = False lowercase_ = False lowercase_ = False def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = FlaxRegNetModelTester(self) __UpperCAmelCase : str = ConfigTester(self , config_class=UpperCamelCase_ , has_text_modality=UpperCamelCase_) def a_ ( self : Dict): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a_ ( self : Tuple): """simple docstring""" return def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase_) @unittest.skip(reason="RegNet does not use inputs_embeds") def a_ ( self : Union[str, Any]): """simple docstring""" pass @unittest.skip(reason="RegNet does not support input and output embeddings") def a_ ( self : Optional[int]): """simple docstring""" pass def a_ ( self : str): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : int = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[int] = inspect.signature(model.__call__) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : Any = [*signature.parameters.keys()] __UpperCAmelCase : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase_) def a_ ( self : int): """simple docstring""" def check_hidden_states_output(UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Union[str, Any]): __UpperCAmelCase : Union[str, Any] = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = model(**self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCAmelCase : str = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase_) , expected_num_stages + 1) __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : List[str] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase : Optional[int] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): __UpperCAmelCase : List[Any] = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[int] = model_class(UpperCamelCase_) @jax.jit def model_jitted(UpperCamelCase_ : int , **UpperCamelCase_ : Optional[int]): return model(pixel_values=UpperCamelCase_ , **UpperCamelCase_) with self.subTest("JIT Enabled"): __UpperCAmelCase : Optional[Any] = model_jitted(**UpperCamelCase_).to_tuple() with self.subTest("JIT Disabled"): with jax.disable_jit(): __UpperCAmelCase : Dict = model_jitted(**UpperCamelCase_).to_tuple() self.assertEqual(len(UpperCamelCase_) , len(UpperCamelCase_)) for jitted_output, output in zip(UpperCamelCase_ , UpperCamelCase_): self.assertEqual(jitted_output.shape , output.shape) def _UpperCamelCase ( ) -> Any: """simple docstring""" __UpperCAmelCase : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_flax class a__ ( unittest.TestCase ): @cached_property def a_ ( self : Optional[int]): """simple docstring""" return AutoImageProcessor.from_pretrained("facebook/regnet-y-040") if is_vision_available() else None @slow def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Any = FlaxRegNetForImageClassification.from_pretrained("facebook/regnet-y-040") __UpperCAmelCase : Dict = self.default_image_processor __UpperCAmelCase : str = prepare_img() __UpperCAmelCase : int = image_processor(images=UpperCamelCase_ , return_tensors="np") __UpperCAmelCase : Dict = model(**UpperCamelCase_) # verify the logits __UpperCAmelCase : Dict = (1, 1000) self.assertEqual(outputs.logits.shape , UpperCamelCase_) __UpperCAmelCase : Any = jnp.array([-0.4180, -1.5051, -3.4836]) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , UpperCamelCase_ , atol=1e-4))
77
1
"""simple docstring""" import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch A = """sshleifer/bart-tiny-random""" A = """patrickvonplaten/t5-tiny-random""" @require_torch class a__ ( unittest.TestCase ): @cached_property def a_ ( self : Optional[int]): """simple docstring""" return AutoConfig.from_pretrained(UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase , *__UpperCAmelCase : Any = create_student_by_copying_alternating_layers(UpperCamelCase_ , tempfile.mkdtemp() , e=1 , d=1) self.assertEqual(student.config.num_hidden_layers , 1) def a_ ( self : int): """simple docstring""" __UpperCAmelCase , *__UpperCAmelCase : str = create_student_by_copying_alternating_layers(UpperCamelCase_ , tempfile.mkdtemp() , e=1 , d=UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase , *__UpperCAmelCase : List[Any] = create_student_by_copying_alternating_layers(UpperCamelCase_ , tempfile.mkdtemp() , e=1 , d=UpperCamelCase_) self.assertEqual(student.config.encoder_layers , 1) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase , *__UpperCAmelCase : int = create_student_by_copying_alternating_layers(UpperCamelCase_ , tempfile.mkdtemp() , e=1 , d=1) self.assertEqual(student.config.encoder_layers , 1) self.assertEqual(student.config.decoder_layers , 1) def a_ ( self : Optional[Any]): """simple docstring""" with self.assertRaises(UpperCamelCase_): create_student_by_copying_alternating_layers(UpperCamelCase_ , tempfile.mkdtemp() , e=UpperCamelCase_ , d=UpperCamelCase_)
77
"""simple docstring""" from scipy.stats import spearmanr import datasets A = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ A = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ A = r"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def a_ ( self : Any): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float"), "references": datasets.Value("float"), }) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def a_ ( self : Union[str, Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : int=False): """simple docstring""" __UpperCAmelCase : List[str] = spearmanr(UpperCamelCase_ , UpperCamelCase_) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
77
1
"""simple docstring""" from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class a__ ( yaml.SafeLoader ): def a_ ( self : Any , UpperCamelCase_ : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[str] = [self.constructed_objects[key_node] for key_node, _ in node.value] __UpperCAmelCase : List[str] = [tuple(UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else key for key in keys] __UpperCAmelCase : List[Any] = Counter(UpperCamelCase_) __UpperCAmelCase : int = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"Got duplicate yaml keys: {duplicate_keys}") def a_ ( self : int , UpperCamelCase_ : List[Any] , UpperCamelCase_ : List[Any]=False): """simple docstring""" __UpperCAmelCase : Tuple = super().construct_mapping(UpperCamelCase_ , deep=UpperCamelCase_) self._check_no_duplicates_on_constructed_node(UpperCamelCase_) return mapping def _UpperCamelCase ( UpperCamelCase ) -> Tuple[Optional[str], str]: """simple docstring""" __UpperCAmelCase : Dict = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: __UpperCAmelCase : int = full_content[1:].index("---" ) + 1 __UpperCAmelCase : Optional[Any] = "\n".join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(UpperCamelCase ) class a__ ( __magic_name__ ): # class attributes lowercase_ = {"train_eval_index"} # train-eval-index in the YAML metadata @classmethod def a_ ( cls : Union[str, Any] , UpperCamelCase_ : Path): """simple docstring""" with open(UpperCamelCase_ , encoding="utf-8") as readme_file: __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = _split_yaml_from_readme(readme_file.read()) if yaml_string is not None: return cls.from_yaml_string(UpperCamelCase_) else: return cls() def a_ ( self : int , UpperCamelCase_ : Path): """simple docstring""" if path.exists(): with open(UpperCamelCase_ , encoding="utf-8") as readme_file: __UpperCAmelCase : Union[str, Any] = readme_file.read() else: __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : int = self._to_readme(UpperCamelCase_) with open(UpperCamelCase_ , "w" , encoding="utf-8") as readme_file: readme_file.write(UpperCamelCase_) def a_ ( self : Optional[int] , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if readme_content is not None: __UpperCAmelCase , __UpperCAmelCase : List[str] = _split_yaml_from_readme(UpperCamelCase_) __UpperCAmelCase : Dict = "---\n" + self.to_yaml_string() + "---\n" + content else: __UpperCAmelCase : Dict = "---\n" + self.to_yaml_string() + "---\n" return full_content @classmethod def a_ ( cls : Any , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : Dict = yaml.load(UpperCamelCase_ , Loader=_NoDuplicateSafeLoader) or {} # Convert the YAML keys to DatasetMetadata fields __UpperCAmelCase : Dict = { (key.replace("-" , "_") if key.replace("-" , "_") in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" return yaml.safe_dump( { (key.replace("_" , "-") if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=UpperCamelCase_ , allow_unicode=UpperCamelCase_ , encoding="utf-8" , ).decode("utf-8") A = { """image-classification""": [], """translation""": [], """image-segmentation""": [], """fill-mask""": [], """automatic-speech-recognition""": [], """token-classification""": [], """sentence-similarity""": [], """audio-classification""": [], """question-answering""": [], """summarization""": [], """zero-shot-classification""": [], """table-to-text""": [], """feature-extraction""": [], """other""": [], """multiple-choice""": [], """text-classification""": [], """text-to-image""": [], """text2text-generation""": [], """zero-shot-image-classification""": [], """tabular-classification""": [], """tabular-regression""": [], """image-to-image""": [], """tabular-to-text""": [], """unconditional-image-generation""": [], """text-retrieval""": [], """text-to-speech""": [], """object-detection""": [], """audio-to-audio""": [], """text-generation""": [], """conversational""": [], """table-question-answering""": [], """visual-question-answering""": [], """image-to-text""": [], """reinforcement-learning""": [], """voice-activity-detection""": [], """time-series-forecasting""": [], """document-question-answering""": [], } if __name__ == "__main__": from argparse import ArgumentParser A = ArgumentParser(usage="""Validate the yaml metadata block of a README.md file.""") ap.add_argument("""readme_filepath""") A = ap.parse_args() A = Path(args.readme_filepath) A = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
77
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) A = {"""vocab_file""": """spiece.model"""} A = { """vocab_file""": { """bert_for_seq_generation""": ( """https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model""" ), } } A = {"""bert_for_seq_generation""": 512} class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = [] lowercase_ = ["input_ids", "attention_mask"] def __init__( self : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : List[str]="<s>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Optional[int]="<unk>" , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : List[Any]="<::::>" , UpperCamelCase_ : Optional[Dict[str, Any]] = None , **UpperCamelCase_ : List[Any] , ): """simple docstring""" __UpperCAmelCase : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) __UpperCAmelCase : Dict = vocab_file __UpperCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(UpperCamelCase_) @property def a_ ( self : List[str]): """simple docstring""" return self.sp_model.get_piece_size() def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : int = {self.convert_ids_to_tokens(UpperCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self : int): """simple docstring""" __UpperCAmelCase : Optional[int] = self.__dict__.copy() __UpperCAmelCase : List[Any] = None return state def __setstate__( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): __UpperCAmelCase : List[Any] = {} __UpperCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def a_ ( self : Any , UpperCamelCase_ : str): """simple docstring""" return self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_) def a_ ( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" return self.sp_model.piece_to_id(UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : int = self.sp_model.IdToPiece(UpperCamelCase_) return token def a_ ( self : Dict , UpperCamelCase_ : Optional[Any]): """simple docstring""" __UpperCAmelCase : int = [] __UpperCAmelCase : Tuple = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCamelCase_) + token __UpperCAmelCase : List[Any] = [] else: current_sub_tokens.append(UpperCamelCase_) out_string += self.sp_model.decode(UpperCamelCase_) return out_string.strip() def a_ ( self : List[str] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : Tuple = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , UpperCamelCase_) elif not os.path.isfile(self.vocab_file): with open(UpperCamelCase_ , "wb") as fi: __UpperCAmelCase : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class a__ : def __init__( self : str , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : Optional[int]=13 , UpperCamelCase_ : Any=7 , UpperCamelCase_ : Union[str, Any]=True , UpperCamelCase_ : Union[str, Any]=True , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : int=True , UpperCamelCase_ : List[Any]=99 , UpperCamelCase_ : Dict=32 , UpperCamelCase_ : List[str]=5 , UpperCamelCase_ : Optional[Any]=4 , UpperCamelCase_ : Union[str, Any]=37 , UpperCamelCase_ : int="gelu" , UpperCamelCase_ : Optional[Any]=0.1 , UpperCamelCase_ : str=0.1 , UpperCamelCase_ : str=512 , UpperCamelCase_ : Tuple=16 , UpperCamelCase_ : int=2 , UpperCamelCase_ : List[Any]=0.02 , UpperCamelCase_ : Union[str, Any]=3 , UpperCamelCase_ : str=4 , UpperCamelCase_ : Optional[Any]=None , ): """simple docstring""" __UpperCAmelCase : List[str] = parent __UpperCAmelCase : Optional[Any] = batch_size __UpperCAmelCase : int = seq_length __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : Optional[int] = use_input_mask __UpperCAmelCase : Optional[Any] = use_token_type_ids __UpperCAmelCase : Optional[Any] = use_labels __UpperCAmelCase : Dict = vocab_size __UpperCAmelCase : Dict = hidden_size __UpperCAmelCase : Union[str, Any] = num_hidden_layers __UpperCAmelCase : Any = num_attention_heads __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : Dict = hidden_act __UpperCAmelCase : Any = hidden_dropout_prob __UpperCAmelCase : Dict = attention_probs_dropout_prob __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : str = type_vocab_size __UpperCAmelCase : Optional[Any] = type_sequence_label_size __UpperCAmelCase : Any = initializer_range __UpperCAmelCase : Any = num_labels __UpperCAmelCase : List[str] = num_choices __UpperCAmelCase : Optional[Any] = scope def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __UpperCAmelCase : Union[str, Any] = None if self.use_input_mask: __UpperCAmelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length]) __UpperCAmelCase : Optional[Any] = None if self.use_token_type_ids: __UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __UpperCAmelCase : Optional[int] = None __UpperCAmelCase : Any = None __UpperCAmelCase : Any = None if self.use_labels: __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __UpperCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __UpperCAmelCase : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices) __UpperCAmelCase : Optional[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a_ ( self : str): """simple docstring""" return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCamelCase_ , initializer_range=self.initializer_range , ) def a_ ( self : Any , UpperCamelCase_ : str , UpperCamelCase_ : Any , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Dict , UpperCamelCase_ : Any , UpperCamelCase_ : Dict): """simple docstring""" __UpperCAmelCase : Optional[Any] = LlamaModel(config=UpperCamelCase_) model.to(UpperCamelCase_) model.eval() __UpperCAmelCase : List[Any] = model(UpperCamelCase_ , attention_mask=UpperCamelCase_) __UpperCAmelCase : Dict = model(UpperCamelCase_) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def a_ ( self : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Dict , UpperCamelCase_ : str , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Dict , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : str , ): """simple docstring""" __UpperCAmelCase : Dict = True __UpperCAmelCase : List[str] = LlamaModel(UpperCamelCase_) model.to(UpperCamelCase_) model.eval() __UpperCAmelCase : Any = model( UpperCamelCase_ , attention_mask=UpperCamelCase_ , encoder_hidden_states=UpperCamelCase_ , encoder_attention_mask=UpperCamelCase_ , ) __UpperCAmelCase : List[Any] = model( UpperCamelCase_ , attention_mask=UpperCamelCase_ , encoder_hidden_states=UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = model(UpperCamelCase_ , attention_mask=UpperCamelCase_) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def a_ ( self : str , UpperCamelCase_ : List[str] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : Tuple , UpperCamelCase_ : int , UpperCamelCase_ : List[Any] , UpperCamelCase_ : int , ): """simple docstring""" __UpperCAmelCase : List[Any] = LlamaForCausalLM(config=UpperCamelCase_) model.to(UpperCamelCase_) model.eval() __UpperCAmelCase : str = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ , labels=UpperCamelCase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def a_ ( self : List[str] , UpperCamelCase_ : Dict , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : Any , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple , UpperCamelCase_ : Any , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : int , ): """simple docstring""" __UpperCAmelCase : Dict = True __UpperCAmelCase : List[str] = True __UpperCAmelCase : List[str] = LlamaForCausalLM(config=UpperCamelCase_) model.to(UpperCamelCase_) model.eval() # first forward pass __UpperCAmelCase : Union[str, Any] = model( UpperCamelCase_ , attention_mask=UpperCamelCase_ , encoder_hidden_states=UpperCamelCase_ , encoder_attention_mask=UpperCamelCase_ , use_cache=UpperCamelCase_ , ) __UpperCAmelCase : Optional[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __UpperCAmelCase : List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size) __UpperCAmelCase : Dict = ids_tensor((self.batch_size, 3) , vocab_size=2) # append to next input_ids and __UpperCAmelCase : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1) __UpperCAmelCase : Optional[int] = torch.cat([input_mask, next_mask] , dim=-1) __UpperCAmelCase : Tuple = model( UpperCamelCase_ , attention_mask=UpperCamelCase_ , encoder_hidden_states=UpperCamelCase_ , encoder_attention_mask=UpperCamelCase_ , output_hidden_states=UpperCamelCase_ , )["hidden_states"][0] __UpperCAmelCase : int = model( UpperCamelCase_ , attention_mask=UpperCamelCase_ , encoder_hidden_states=UpperCamelCase_ , encoder_attention_mask=UpperCamelCase_ , past_key_values=UpperCamelCase_ , output_hidden_states=UpperCamelCase_ , )["hidden_states"][0] # select random slice __UpperCAmelCase : List[str] = ids_tensor((1,) , output_from_past.shape[-1]).item() __UpperCAmelCase : Tuple = output_from_no_past[:, -3:, random_slice_idx].detach() __UpperCAmelCase : Tuple = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-3)) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : List[Any] = self.prepare_config_and_inputs() ( ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ) : Any = config_and_inputs __UpperCAmelCase : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class a__ ( __magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase_ = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () lowercase_ = (LlamaForCausalLM,) if is_torch_available() else () lowercase_ = ( { "feature-extraction": LlamaModel, "text-classification": LlamaForSequenceClassification, "text-generation": LlamaForCausalLM, "zero-shot": LlamaForSequenceClassification, } if is_torch_available() else {} ) lowercase_ = False lowercase_ = False def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[int] = LlamaModelTester(self) __UpperCAmelCase : str = ConfigTester(self , config_class=UpperCamelCase_ , hidden_size=37) def a_ ( self : Union[str, Any]): """simple docstring""" self.config_tester.run_common_tests() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __UpperCAmelCase : Optional[Any] = type self.model_tester.create_and_check_model(*UpperCamelCase_) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase : Optional[int] = 3 __UpperCAmelCase : List[Any] = input_dict["input_ids"] __UpperCAmelCase : Tuple = input_ids.ne(1).to(UpperCamelCase_) __UpperCAmelCase : List[str] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) __UpperCAmelCase : Optional[Any] = LlamaForSequenceClassification(UpperCamelCase_) model.to(UpperCamelCase_) model.eval() __UpperCAmelCase : Dict = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ , labels=UpperCamelCase_) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase : Union[str, Any] = 3 __UpperCAmelCase : Optional[Any] = "single_label_classification" __UpperCAmelCase : str = input_dict["input_ids"] __UpperCAmelCase : List[str] = input_ids.ne(1).to(UpperCamelCase_) __UpperCAmelCase : List[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) __UpperCAmelCase : str = LlamaForSequenceClassification(UpperCamelCase_) model.to(UpperCamelCase_) model.eval() __UpperCAmelCase : Any = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ , labels=UpperCamelCase_) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase : Optional[int] = 3 __UpperCAmelCase : Tuple = "multi_label_classification" __UpperCAmelCase : Union[str, Any] = input_dict["input_ids"] __UpperCAmelCase : List[str] = input_ids.ne(1).to(UpperCamelCase_) __UpperCAmelCase : List[str] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size).to(torch.float) __UpperCAmelCase : Any = LlamaForSequenceClassification(UpperCamelCase_) model.to(UpperCamelCase_) model.eval() __UpperCAmelCase : Union[str, Any] = model(UpperCamelCase_ , attention_mask=UpperCamelCase_ , labels=UpperCamelCase_) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) @unittest.skip("LLaMA buffers include complex numbers, which breaks this test") def a_ ( self : Dict): """simple docstring""" pass @parameterized.expand([("linear",), ("dynamic",)]) def a_ ( self : int , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase : str = ids_tensor([1, 10] , config.vocab_size) __UpperCAmelCase : int = ids_tensor([1, int(config.max_position_embeddings * 1.5)] , config.vocab_size) set_seed(42) # Fixed seed at init time so the two models get the same random weights __UpperCAmelCase : Any = LlamaModel(UpperCamelCase_) original_model.to(UpperCamelCase_) original_model.eval() __UpperCAmelCase : Any = original_model(UpperCamelCase_).last_hidden_state __UpperCAmelCase : Tuple = original_model(UpperCamelCase_).last_hidden_state set_seed(42) # Fixed seed at init time so the two models get the same random weights __UpperCAmelCase : str = {"type": scaling_type, "factor": 10.0} __UpperCAmelCase : Union[str, Any] = LlamaModel(UpperCamelCase_) scaled_model.to(UpperCamelCase_) scaled_model.eval() __UpperCAmelCase : Optional[Any] = scaled_model(UpperCamelCase_).last_hidden_state __UpperCAmelCase : Tuple = scaled_model(UpperCamelCase_).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-5)) else: self.assertFalse(torch.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-5)) # The output should be different for long inputs self.assertFalse(torch.allclose(UpperCamelCase_ , UpperCamelCase_ , atol=1e-5)) @require_torch class a__ ( unittest.TestCase ): @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!") @slow def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[int] = [1, 306, 4658, 278, 6593, 310, 2834, 338] __UpperCAmelCase : List[str] = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf" , device_map="auto") __UpperCAmelCase : Optional[Any] = model(torch.tensor([input_ids])) # Expected mean on dim = -1 __UpperCAmelCase : Union[str, Any] = torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]]) torch.testing.assert_close(out.mean(-1) , UpperCamelCase_ , atol=1e-2 , rtol=1e-2) # slicing logits[0, 0, 0:30] # fmt: off __UpperCAmelCase : Dict = torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,]) # fmt: on torch.testing.assert_close(out[0, 0, :30] , UpperCamelCase_ , atol=1e-5 , rtol=1e-5) @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!") @slow def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[Any] = [1, 306, 4658, 278, 6593, 310, 2834, 338] __UpperCAmelCase : List[Any] = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-13b-hf" , device_map="auto") __UpperCAmelCase : Any = model(torch.tensor(UpperCamelCase_)) # Expected mean on dim = -1 __UpperCAmelCase : int = torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]]) torch.testing.assert_close(out.mean(-1) , UpperCamelCase_ , atol=1e-2 , rtol=1e-2) # slicing logits[0, 0, 0:30] # fmt: off __UpperCAmelCase : str = torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273]) # fmt: on torch.testing.assert_close(out[0, 0, :30] , UpperCamelCase_ , atol=1e-5 , rtol=1e-5) @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!") @slow def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : str = [1, 306, 4658, 278, 6593, 310, 2834, 338] __UpperCAmelCase : Any = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-13b-chat-hf" , device_map="auto") __UpperCAmelCase : Optional[Any] = model(torch.tensor(UpperCamelCase_)) # Expected mean on dim = -1 __UpperCAmelCase : int = torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]]) torch.testing.assert_close(out.mean(-1) , UpperCamelCase_ , atol=1e-2 , rtol=1e-2) # slicing logits[0, 0, 0:30] # fmt: off __UpperCAmelCase : Optional[Any] = torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513]) # fmt: on torch.testing.assert_close(out.mean(-1) , UpperCamelCase_ , atol=1e-2 , rtol=1e-2) @unittest.skip( "Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test") @slow def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = [1, 306, 4658, 278, 6593, 310, 2834, 338] __UpperCAmelCase : str = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-70b-hf" , device_map="auto") __UpperCAmelCase : Tuple = model(torch.tensor(UpperCamelCase_)) __UpperCAmelCase : Optional[int] = torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa) torch.testing.assert_close(out.mean(-1) , UpperCamelCase_ , atol=1e-2 , rtol=1e-2) # fmt: off __UpperCAmelCase : List[Any] = torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312]) # fmt: on torch.testing.assert_close(out[0, 0, :30] , UpperCamelCase_ , atol=1e-5 , rtol=1e-5) @unittest.skip("Model is curently gated") @slow def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Dict = "Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the \"princi" __UpperCAmelCase : int = "Simply put, the theory of relativity states that " __UpperCAmelCase : Union[str, Any] = LlamaTokenizer.from_pretrained("meta-llama/Llama-2-13b-chat-hf") __UpperCAmelCase : str = tokenizer.encode(UpperCamelCase_ , return_tensors="pt") __UpperCAmelCase : int = LlamaForCausalLM.from_pretrained( "meta-llama/Llama-2-13b-chat-hf" , device_map="sequential" , use_safetensors=UpperCamelCase_) # greedy generation outputs __UpperCAmelCase : int = model.generate(UpperCamelCase_ , max_new_tokens=64 , top_p=UpperCamelCase_ , temperature=1 , do_sample=UpperCamelCase_) __UpperCAmelCase : List[Any] = tokenizer.decode(generated_ids[0] , skip_special_tokens=UpperCamelCase_) self.assertEqual(UpperCamelCase_ , UpperCamelCase_)
77
"""simple docstring""" import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed A = """true""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=16 ) -> Tuple: """simple docstring""" set_seed(42 ) __UpperCAmelCase : Dict = RegressionModel() __UpperCAmelCase : Optional[Any] = deepcopy(UpperCamelCase ) __UpperCAmelCase : Any = RegressionDataset(length=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = DataLoader(UpperCamelCase , batch_size=UpperCamelCase ) model.to(accelerator.device ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return model, ddp_model, dataloader def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=False ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) __UpperCAmelCase : Dict = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(UpperCamelCase ): __UpperCAmelCase : Dict = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=UpperCamelCase , max_length=UpperCamelCase ) return outputs with accelerator.main_process_first(): __UpperCAmelCase : str = dataset.map( UpperCamelCase , batched=UpperCamelCase , remove_columns=["idx", "sentence1", "sentence2"] , ) __UpperCAmelCase : List[Any] = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(UpperCamelCase ): if use_longest: return tokenizer.pad(UpperCamelCase , padding="longest" , return_tensors="pt" ) return tokenizer.pad(UpperCamelCase , padding="max_length" , max_length=128 , return_tensors="pt" ) return DataLoader(UpperCamelCase , shuffle=UpperCamelCase , collate_fn=UpperCamelCase , batch_size=16 ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[Any] = Accelerator(dispatch_batches=UpperCamelCase , split_batches=UpperCamelCase ) __UpperCAmelCase : int = get_dataloader(UpperCamelCase , not dispatch_batches ) __UpperCAmelCase : Any = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Dict = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Dict = [] for batch in dataloader: __UpperCAmelCase , __UpperCAmelCase : int = batch.values() with torch.no_grad(): __UpperCAmelCase : int = model(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : List[str] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = [], [] for logit, targ in logits_and_targets: logits.append(UpperCamelCase ) targs.append(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = torch.cat(UpperCamelCase ), torch.cat(UpperCamelCase ) return logits, targs def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=16 ) -> int: """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = get_basic_setup(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = generate_predictions(UpperCamelCase , UpperCamelCase , UpperCamelCase ) assert ( len(UpperCamelCase ) == num_samples ), f"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(UpperCamelCase )}" def _UpperCamelCase ( UpperCamelCase = False , UpperCamelCase = False ) -> List[str]: """simple docstring""" __UpperCAmelCase : List[str] = evaluate.load("glue" , "mrpc" ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = get_mrpc_setup(UpperCamelCase , UpperCamelCase ) # First do baseline __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = setup["no"] model.to(UpperCamelCase ) model.eval() for batch in dataloader: batch.to(UpperCamelCase ) with torch.inference_mode(): __UpperCAmelCase : List[str] = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=UpperCamelCase , references=batch["labels"] ) __UpperCAmelCase : str = metric.compute() # Then do distributed __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): __UpperCAmelCase : Any = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) __UpperCAmelCase : Union[str, Any] = batch["labels"] __UpperCAmelCase , __UpperCAmelCase : Any = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=UpperCamelCase , references=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def _UpperCamelCase ( ) -> List[Any]: """simple docstring""" __UpperCAmelCase : Dict = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(UpperCamelCase , UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __UpperCAmelCase : Union[str, Any] = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(UpperCamelCase , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) __UpperCAmelCase : Any = Accelerator() test_torch_metrics(UpperCamelCase , 512 ) accelerator.state._reset_state() def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) A = { """configuration_perceiver""": ["""PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PerceiverConfig""", """PerceiverOnnxConfig"""], """tokenization_perceiver""": ["""PerceiverTokenizer"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""PerceiverFeatureExtractor"""] A = ["""PerceiverImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST""", """PerceiverForImageClassificationConvProcessing""", """PerceiverForImageClassificationFourier""", """PerceiverForImageClassificationLearned""", """PerceiverForMaskedLM""", """PerceiverForMultimodalAutoencoding""", """PerceiverForOpticalFlow""", """PerceiverForSequenceClassification""", """PerceiverLayer""", """PerceiverModel""", """PerceiverPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" import math def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = 0 , UpperCamelCase = 0 ) -> list: """simple docstring""" __UpperCAmelCase : Union[str, Any] = end or len(UpperCamelCase ) for i in range(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[Any] = i __UpperCAmelCase : Any = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: __UpperCAmelCase : Dict = array[temp_index - 1] temp_index -= 1 __UpperCAmelCase : str = temp_index_value return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: # Max Heap """simple docstring""" __UpperCAmelCase : Optional[Any] = index __UpperCAmelCase : List[str] = 2 * index + 1 # Left Node __UpperCAmelCase : Union[str, Any] = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: __UpperCAmelCase : Tuple = left_index if right_index < heap_size and array[largest] < array[right_index]: __UpperCAmelCase : int = right_index if largest != index: __UpperCAmelCase , __UpperCAmelCase : List[str] = array[largest], array[index] heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" __UpperCAmelCase : List[Any] = len(UpperCamelCase ) for i in range(n // 2 , -1 , -1 ): heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) for i in range(n - 1 , 0 , -1 ): __UpperCAmelCase , __UpperCAmelCase : int = array[0], array[i] heapify(UpperCamelCase , 0 , UpperCamelCase ) return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Optional[Any] = low __UpperCAmelCase : List[str] = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i __UpperCAmelCase , __UpperCAmelCase : Optional[int] = array[j], array[i] i += 1 def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" if len(UpperCamelCase ) == 0: return array __UpperCAmelCase : Optional[int] = 2 * math.ceil(math.loga(len(UpperCamelCase ) ) ) __UpperCAmelCase : List[Any] = 16 return intro_sort(UpperCamelCase , 0 , len(UpperCamelCase ) , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> list: """simple docstring""" while end - start > size_threshold: if max_depth == 0: return heap_sort(UpperCamelCase ) max_depth -= 1 __UpperCAmelCase : List[Any] = median_of_a(UpperCamelCase , UpperCamelCase , start + ((end - start) // 2) + 1 , end - 1 ) __UpperCAmelCase : Union[str, Any] = partition(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) intro_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[Any] = p return insertion_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() A = input("""Enter numbers separated by a comma : """).strip() A = [float(item) for item in user_input.split(""",""")] print(sort(unsorted))
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" return " ".join( "".join(word[::-1] ) if len(UpperCamelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
77
"""simple docstring""" import numpy as np from PIL import Image def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : str = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Any = 0 __UpperCAmelCase : Dict = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Tuple = 0 # compute the shape of the output matrix __UpperCAmelCase : Optional[int] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __UpperCAmelCase : List[str] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __UpperCAmelCase : str = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : int = 0 __UpperCAmelCase : int = 0 return updated_arr def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : List[str] = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Any = 0 # compute the shape of the output matrix __UpperCAmelCase : Tuple = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __UpperCAmelCase : str = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __UpperCAmelCase : Tuple = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Optional[Any] = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name="""avgpooling""", verbose=True) # Loading the image A = Image.open("""path_to_image""") # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
77
1
"""simple docstring""" import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_led import LEDTokenizer A = logging.get_logger(__name__) A = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} A = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } A = { """allenai/led-base-16384""": 16_384, } class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = LEDTokenizer lowercase_ = ["input_ids", "attention_mask"] def __init__( self : Union[str, Any] , UpperCamelCase_ : Any=None , UpperCamelCase_ : List[Any]=None , UpperCamelCase_ : Dict=None , UpperCamelCase_ : str="replace" , UpperCamelCase_ : List[Any]="<s>" , UpperCamelCase_ : Dict="</s>" , UpperCamelCase_ : Tuple="</s>" , UpperCamelCase_ : List[Any]="<s>" , UpperCamelCase_ : Optional[int]="<unk>" , UpperCamelCase_ : Any="<pad>" , UpperCamelCase_ : List[Any]="<mask>" , UpperCamelCase_ : Optional[Any]=False , UpperCamelCase_ : List[Any]=True , **UpperCamelCase_ : List[str] , ): """simple docstring""" super().__init__( UpperCamelCase_ , UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , errors=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , add_prefix_space=UpperCamelCase_ , trim_offsets=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Optional[Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get("add_prefix_space" , UpperCamelCase_) != add_prefix_space: __UpperCAmelCase : str = getattr(UpperCamelCase_ , pre_tok_state.pop("type")) __UpperCAmelCase : Any = add_prefix_space __UpperCAmelCase : Union[str, Any] = pre_tok_class(**UpperCamelCase_) __UpperCAmelCase : List[str] = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` __UpperCAmelCase : Optional[int] = "post_processor" __UpperCAmelCase : Optional[int] = getattr(self.backend_tokenizer , UpperCamelCase_ , UpperCamelCase_) if tokenizer_component_instance: __UpperCAmelCase : List[str] = json.loads(tokenizer_component_instance.__getstate__()) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __UpperCAmelCase : Dict = tuple(state["sep"]) if "cls" in state: __UpperCAmelCase : Dict = tuple(state["cls"]) __UpperCAmelCase : Union[str, Any] = False if state.get("add_prefix_space" , UpperCamelCase_) != add_prefix_space: __UpperCAmelCase : Optional[Any] = add_prefix_space __UpperCAmelCase : List[str] = True if state.get("trim_offsets" , UpperCamelCase_) != trim_offsets: __UpperCAmelCase : Any = trim_offsets __UpperCAmelCase : Optional[Any] = True if changes_to_apply: __UpperCAmelCase : Any = getattr(UpperCamelCase_ , state.pop("type")) __UpperCAmelCase : str = component_class(**UpperCamelCase_) setattr(self.backend_tokenizer , UpperCamelCase_ , UpperCamelCase_) @property # Copied from transformers.models.bart.tokenization_bart_fast.BartTokenizerFast.mask_token with BART->LED def a_ ( self : List[Any]): """simple docstring""" if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet.") return None return str(self._mask_token) @mask_token.setter def a_ ( self : List[str] , UpperCamelCase_ : List[str]): """simple docstring""" __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else value __UpperCAmelCase : Any = value def a_ ( self : Tuple , *UpperCamelCase_ : Union[str, Any] , **UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = kwargs.get("is_split_into_words" , UpperCamelCase_) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs.") return super()._batch_encode_plus(*UpperCamelCase_ , **UpperCamelCase_) def a_ ( self : Optional[Any] , *UpperCamelCase_ : Tuple , **UpperCamelCase_ : List[Any]): """simple docstring""" __UpperCAmelCase : List[Any] = kwargs.get("is_split_into_words" , UpperCamelCase_) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " "to use it with pretokenized inputs.") return super()._encode_plus(*UpperCamelCase_ , **UpperCamelCase_) def a_ ( self : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" __UpperCAmelCase : int = self._tokenizer.model.save(UpperCamelCase_ , name=UpperCamelCase_) return tuple(UpperCamelCase_) def a_ ( self : Union[str, Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Optional[Any]=None): """simple docstring""" __UpperCAmelCase : List[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def a_ ( self : str , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : List[str] = [self.sep_token_id] __UpperCAmelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] def a_ ( self : Optional[int] , UpperCamelCase_ : Union[Dict[str, EncodedInput], BatchEncoding] , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[bool] = None , ): """simple docstring""" __UpperCAmelCase : Dict = super()._pad( encoded_inputs=UpperCamelCase_ , max_length=UpperCamelCase_ , padding_strategy=UpperCamelCase_ , pad_to_multiple_of=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , ) # Load from model defaults if return_attention_mask is None: __UpperCAmelCase : int = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: __UpperCAmelCase : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. __UpperCAmelCase : str = len(encoded_inputs["global_attention_mask"]) != len(UpperCamelCase_) if needs_to_be_padded: __UpperCAmelCase : List[Any] = len(UpperCamelCase_) - len(encoded_inputs["global_attention_mask"]) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` __UpperCAmelCase : Any = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": __UpperCAmelCase : Union[str, Any] = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side)) return encoded_inputs
77
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: A = None A = logging.get_logger(__name__) A = """▁""" A = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} A = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } A = { """google/pegasus-xsum""": 512, } class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = PegasusTokenizer lowercase_ = ["input_ids", "attention_mask"] def __init__( self : str , UpperCamelCase_ : str=None , UpperCamelCase_ : Optional[int]=None , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Any="<unk>" , UpperCamelCase_ : Tuple="<mask_2>" , UpperCamelCase_ : Any="<mask_1>" , UpperCamelCase_ : Tuple=None , UpperCamelCase_ : str=103 , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" __UpperCAmelCase : Optional[int] = offset if additional_special_tokens is not None: if not isinstance(UpperCamelCase_ , UpperCamelCase_): raise TypeError( F"additional_special_tokens should be of type {type(UpperCamelCase_)}, but is" F" {type(UpperCamelCase_)}") __UpperCAmelCase : Any = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"<unk_{i}>" for i in range(len(UpperCamelCase_) , self.offset - 1) ] if len(set(UpperCamelCase_)) != len(UpperCamelCase_): raise ValueError( "Please make sure that the provided additional_special_tokens do not contain an incorrectly" F" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.") __UpperCAmelCase : str = additional_special_tokens_extended else: __UpperCAmelCase : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"<unk_{i}>" for i in range(2 , self.offset)] super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , pad_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , mask_token_sent=UpperCamelCase_ , offset=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Optional[int] = vocab_file __UpperCAmelCase : List[str] = False if not self.vocab_file else True def a_ ( self : Union[str, Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : int = set(self.all_special_ids) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens) + 3)): raise ValueError( "There should be 3 special tokens: mask_token, pad_token, and eos_token +" F" {len(self.additional_special_tokens)} additional_special_tokens, but got {all_special_ids}") return [1 if x in all_special_ids else 0 for x in seq] def a_ ( self : Union[str, Any] , UpperCamelCase_ : List , UpperCamelCase_ : Optional[List] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return self._special_token_mask(UpperCamelCase_) elif token_ids_a is None: return self._special_token_mask(UpperCamelCase_) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a) + [1] def a_ ( self : int , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : List[Any]=None): """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a_ ( self : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : List[str] = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_): copyfile(self.vocab_file , UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) __UpperCAmelCase : Optional[Any] = str(bin(UpperCamelCase ) )[2:] # remove the leading "0b" __UpperCAmelCase : int = str(bin(UpperCamelCase ) )[2:] # remove the leading "0b" __UpperCAmelCase : Tuple = max(len(UpperCamelCase ) , len(UpperCamelCase ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(UpperCamelCase ) , b_binary.zfill(UpperCamelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __UpperCAmelCase : Optional[Any] = TapasConfig.from_json_file(UpperCamelCase ) # set absolute/relative position embeddings parameter __UpperCAmelCase : Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __UpperCAmelCase : List[str] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WTQ": # run_task_main.py hparams __UpperCAmelCase : Tuple = 4 __UpperCAmelCase : Any = True # hparam_utils.py hparams __UpperCAmelCase : Union[str, Any] = 0.664694 __UpperCAmelCase : Union[str, Any] = 0.207951 __UpperCAmelCase : int = 0.121194 __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : List[str] = True __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : List[str] = 0.0352513 __UpperCAmelCase : Optional[int] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __UpperCAmelCase : int = 4 __UpperCAmelCase : Optional[int] = False # hparam_utils.py hparams __UpperCAmelCase : int = 36.4519 __UpperCAmelCase : str = 0.903421 __UpperCAmelCase : Dict = 222.088 __UpperCAmelCase : Dict = True __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Tuple = True __UpperCAmelCase : Any = 0.763141 __UpperCAmelCase : Optional[Any] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "TABFACT": __UpperCAmelCase : Union[str, Any] = TapasForSequenceClassification(config=UpperCamelCase ) elif task == "MLM": __UpperCAmelCase : Tuple = TapasForMaskedLM(config=UpperCamelCase ) elif task == "INTERMEDIATE_PRETRAINING": __UpperCAmelCase : List[str] = TapasModel(config=UpperCamelCase ) else: raise ValueError(f"Task {task} not supported." ) print(f"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # Save pytorch-model (weights and configuration) print(f"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(UpperCamelCase ) # Save tokenizer files print(f"Save tokenizer files to {pytorch_dump_path}" ) __UpperCAmelCase : str = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + "vocab.txt" , model_max_length=512 ) tokenizer.save_pretrained(UpperCamelCase ) print("Used relative position embeddings:" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
77
1
"""simple docstring""" import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Tuple: """simple docstring""" __UpperCAmelCase : int = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") __UpperCAmelCase : Optional[int] = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(UpperCamelCase ): os.makedirs(UpperCamelCase ) __UpperCAmelCase : Optional[Any] = model.state_dict() def to_tf_var_name(UpperCamelCase ): for patt, repl in iter(UpperCamelCase ): __UpperCAmelCase : Any = name.replace(UpperCamelCase , UpperCamelCase ) return f"bert/{name}" def create_tf_var(UpperCamelCase , UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : str = tf.dtypes.as_dtype(tensor.dtype ) __UpperCAmelCase : Tuple = tf.get_variable(dtype=UpperCamelCase , shape=tensor.shape , name=UpperCamelCase , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(UpperCamelCase ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: __UpperCAmelCase : Union[str, Any] = to_tf_var_name(UpperCamelCase ) __UpperCAmelCase : List[str] = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): __UpperCAmelCase : Tuple = torch_tensor.T __UpperCAmelCase : str = create_tf_var(tensor=UpperCamelCase , name=UpperCamelCase , session=UpperCamelCase ) tf.keras.backend.set_value(UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Dict = session.run(UpperCamelCase ) print(f"Successfully created {tf_name}: {np.allclose(UpperCamelCase , UpperCamelCase )}" ) __UpperCAmelCase : Optional[int] = tf.train.Saver(tf.trainable_variables() ) saver.save(UpperCamelCase , os.path.join(UpperCamelCase , model_name.replace("-" , "_" ) + ".ckpt" ) ) def _UpperCamelCase ( UpperCamelCase=None ) -> int: """simple docstring""" __UpperCAmelCase : List[Any] = argparse.ArgumentParser() parser.add_argument("--model_name" , type=UpperCamelCase , required=UpperCamelCase , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=UpperCamelCase , default=UpperCamelCase , required=UpperCamelCase , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=UpperCamelCase , required=UpperCamelCase , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=UpperCamelCase , required=UpperCamelCase , help="Directory in which to save tensorflow model" ) __UpperCAmelCase : List[Any] = parser.parse_args(UpperCamelCase ) __UpperCAmelCase : List[str] = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=UpperCamelCase , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
77
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = "cpu" , UpperCamelCase = None ) -> None: """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.load(UpperCamelCase , map_location=UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(UpperCamelCase , torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) __UpperCAmelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __UpperCAmelCase : str = src_path torch.save(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
77
1
"""simple docstring""" from numpy import exp, pi, sqrt def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = 0.0 , UpperCamelCase = 1.0 ) -> int: """simple docstring""" return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": A = pd.read_csv("""sample_data.csv""", header=None) A = df.shape[:1][0] # If you're using some other dataset input the target column A = df.iloc[:, 1:2] A = actual_data.values.reshape(len_data, 1) A = MinMaxScaler().fit_transform(actual_data) A = 10 A = 5 A = 20 A = len_data - periods * look_back A = actual_data[:division] A = actual_data[division - look_back :] A , A = [], [] A , A = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) A = np.array(train_x) A = np.array(test_x) A = np.array([list(i.ravel()) for i in train_y]) A = np.array([list(i.ravel()) for i in test_y]) A = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") A = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) A = model.predict(x_test)
77
1
"""simple docstring""" print((lambda quine: quine % quine)("""print((lambda quine: quine %% quine)(%r))"""))
77
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A = 250_004 A = 250_020 @require_sentencepiece @require_tokenizers class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = MBartTokenizer lowercase_ = MBartTokenizerFast lowercase_ = True lowercase_ = True def a_ ( self : str): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase : Any = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) tokenizer.save_pretrained(self.tmpdirname) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Dict = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer.tokenize("This is a test") self.assertListEqual(UpperCamelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase_) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __UpperCAmelCase : List[Any] = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __UpperCAmelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def a_ ( self : Dict): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __UpperCAmelCase : Dict = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})"): __UpperCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = tempfile.mkdtemp() __UpperCAmelCase : Optional[int] = tokenizer_r.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Any = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) __UpperCAmelCase : Any = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f) self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Tuple = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=True __UpperCAmelCase : Optional[int] = tempfile.mkdtemp() __UpperCAmelCase : Dict = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : int = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=False __UpperCAmelCase : Tuple = tempfile.mkdtemp() __UpperCAmelCase : int = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way __UpperCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : str = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) @require_torch @require_sentencepiece @require_tokenizers class a__ ( unittest.TestCase ): lowercase_ = "facebook/mbart-large-en-ro" lowercase_ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowercase_ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowercase_ = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def a_ ( cls : int): """simple docstring""" __UpperCAmelCase : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en_XX" , tgt_lang="ro_RO") __UpperCAmelCase : Union[str, Any] = 1 return cls def a_ ( self : List[Any]): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ar_AR"] , 250001) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["en_EN"] , 250004) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ro_RO"] , 250020) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" self.assertIn(UpperCamelCase_ , self.tokenizer.all_special_ids) __UpperCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] __UpperCAmelCase : Optional[Any] = self.tokenizer.decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase_) self.assertEqual(UpperCamelCase_ , UpperCamelCase_) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , UpperCamelCase_) __UpperCAmelCase : Tuple = 10 __UpperCAmelCase : List[Any] = self.tokenizer(UpperCamelCase_ , max_length=UpperCamelCase_ , truncation=UpperCamelCase_).input_ids[0] self.assertEqual(ids[-2] , 2) self.assertEqual(ids[-1] , UpperCamelCase_) self.assertEqual(len(UpperCamelCase_) , UpperCamelCase_) def a_ ( self : Any): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"]) , [250026, 250001]) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[str] = tempfile.mkdtemp() __UpperCAmelCase : Union[str, Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase_) __UpperCAmelCase : List[Any] = MBartTokenizer.from_pretrained(UpperCamelCase_) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCamelCase_) @require_torch def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , return_tensors="pt") __UpperCAmelCase : Dict = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=len(self.expected_src_tokens) , return_tensors="pt" , ) __UpperCAmelCase : Tuple = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_) self.assertEqual((2, 14) , batch.input_ids.shape) self.assertEqual((2, 14) , batch.attention_mask.shape) __UpperCAmelCase : List[str] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) self.assertEqual(2 , batch.decoder_input_ids[0, -1]) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , []) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE]) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : List[str] = self.tokenizer(self.src_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=3 , return_tensors="pt") __UpperCAmelCase : Any = self.tokenizer( text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=10 , return_tensors="pt") __UpperCAmelCase : int = targets["input_ids"] __UpperCAmelCase : Any = shift_tokens_right(UpperCamelCase_ , self.tokenizer.pad_token_id) self.assertEqual(batch.input_ids.shape[1] , 3) self.assertEqual(batch.decoder_input_ids.shape[1] , 10) @require_torch def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="ar_AR") self.assertEqual( nested_simplify(UpperCamelCase_) , { # A, test, EOS, en_XX "input_ids": [[62, 3034, 2, 250004]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 250001, } , )
77
1
"""simple docstring""" import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A = logging.get_logger(__name__) A = { """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class a__ ( __magic_name__ ): lowercase_ = "conditional_detr" lowercase_ = ["past_key_values"] lowercase_ = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self : Tuple , UpperCamelCase_ : Optional[Any]=True , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : List[str]=3 , UpperCamelCase_ : int=300 , UpperCamelCase_ : List[Any]=6 , UpperCamelCase_ : Optional[int]=2048 , UpperCamelCase_ : str=8 , UpperCamelCase_ : Tuple=6 , UpperCamelCase_ : int=2048 , UpperCamelCase_ : Optional[Any]=8 , UpperCamelCase_ : Optional[int]=0.0 , UpperCamelCase_ : Tuple=0.0 , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : int="relu" , UpperCamelCase_ : Optional[int]=256 , UpperCamelCase_ : str=0.1 , UpperCamelCase_ : Union[str, Any]=0.0 , UpperCamelCase_ : Any=0.0 , UpperCamelCase_ : Optional[int]=0.02 , UpperCamelCase_ : str=1.0 , UpperCamelCase_ : Any=False , UpperCamelCase_ : Union[str, Any]="sine" , UpperCamelCase_ : Tuple="resnet50" , UpperCamelCase_ : str=True , UpperCamelCase_ : Any=False , UpperCamelCase_ : Dict=2 , UpperCamelCase_ : str=5 , UpperCamelCase_ : Tuple=2 , UpperCamelCase_ : str=1 , UpperCamelCase_ : str=1 , UpperCamelCase_ : Optional[int]=2 , UpperCamelCase_ : Tuple=5 , UpperCamelCase_ : Any=2 , UpperCamelCase_ : Optional[int]=0.25 , **UpperCamelCase_ : List[Any] , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`.") if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.") __UpperCAmelCase : Optional[Any] = CONFIG_MAPPING["resnet"](out_features=["stage4"]) elif isinstance(UpperCamelCase_ , UpperCamelCase_): __UpperCAmelCase : int = backbone_config.get("model_type") __UpperCAmelCase : Dict = CONFIG_MAPPING[backbone_model_type] __UpperCAmelCase : int = config_class.from_dict(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = use_timm_backbone __UpperCAmelCase : Optional[int] = backbone_config __UpperCAmelCase : Union[str, Any] = num_channels __UpperCAmelCase : Optional[int] = num_queries __UpperCAmelCase : List[Any] = d_model __UpperCAmelCase : int = encoder_ffn_dim __UpperCAmelCase : Optional[int] = encoder_layers __UpperCAmelCase : Optional[int] = encoder_attention_heads __UpperCAmelCase : List[Any] = decoder_ffn_dim __UpperCAmelCase : Optional[Any] = decoder_layers __UpperCAmelCase : str = decoder_attention_heads __UpperCAmelCase : Union[str, Any] = dropout __UpperCAmelCase : Optional[int] = attention_dropout __UpperCAmelCase : List[Any] = activation_dropout __UpperCAmelCase : Tuple = activation_function __UpperCAmelCase : List[Any] = init_std __UpperCAmelCase : List[str] = init_xavier_std __UpperCAmelCase : Dict = encoder_layerdrop __UpperCAmelCase : Optional[int] = decoder_layerdrop __UpperCAmelCase : int = encoder_layers __UpperCAmelCase : Optional[int] = auxiliary_loss __UpperCAmelCase : Optional[int] = position_embedding_type __UpperCAmelCase : Optional[int] = backbone __UpperCAmelCase : List[Any] = use_pretrained_backbone __UpperCAmelCase : Optional[Any] = dilation # Hungarian matcher __UpperCAmelCase : List[str] = class_cost __UpperCAmelCase : Optional[int] = bbox_cost __UpperCAmelCase : Dict = giou_cost # Loss coefficients __UpperCAmelCase : Dict = mask_loss_coefficient __UpperCAmelCase : Any = dice_loss_coefficient __UpperCAmelCase : str = cls_loss_coefficient __UpperCAmelCase : int = bbox_loss_coefficient __UpperCAmelCase : str = giou_loss_coefficient __UpperCAmelCase : List[Any] = focal_alpha super().__init__(is_encoder_decoder=UpperCamelCase_ , **UpperCamelCase_) @property def a_ ( self : Union[str, Any]): """simple docstring""" return self.encoder_attention_heads @property def a_ ( self : int): """simple docstring""" return self.d_model def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Tuple = copy.deepcopy(self.__dict__) if self.backbone_config is not None: __UpperCAmelCase : int = self.backbone_config.to_dict() __UpperCAmelCase : Tuple = self.__class__.model_type return output class a__ ( __magic_name__ ): lowercase_ = version.parse("1.11" ) @property def a_ ( self : Optional[Any]): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ]) @property def a_ ( self : Tuple): """simple docstring""" return 1e-5 @property def a_ ( self : Tuple): """simple docstring""" return 12
77
"""simple docstring""" from typing import Any class a__ : def __init__( self : List[str] , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : str = data __UpperCAmelCase : Optional[Any] = None class a__ : def __init__( self : Any): """simple docstring""" __UpperCAmelCase : Optional[int] = None def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.head while temp is not None: print(temp.data , end=" ") __UpperCAmelCase : Tuple = temp.next print() def a_ ( self : int , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : List[str] = Node(UpperCamelCase_) __UpperCAmelCase : str = self.head __UpperCAmelCase : Optional[int] = new_node def a_ ( self : str , UpperCamelCase_ : List[Any] , UpperCamelCase_ : str): """simple docstring""" if node_data_a == node_data_a: return else: __UpperCAmelCase : int = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Tuple = node_a.next __UpperCAmelCase : List[Any] = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Optional[Any] = node_a.next if node_a is None or node_a is None: return __UpperCAmelCase , __UpperCAmelCase : Any = node_a.data, node_a.data if __name__ == "__main__": A = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
77
1
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=__magic_name__ ) class a__ ( __magic_name__ ): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowercase_ = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True} ) lowercase_ = Features({"text": Value("string" )} ) lowercase_ = Features({"labels": ClassLabel} ) lowercase_ = "text" lowercase_ = "labels" def a_ ( self : List[Any] , UpperCamelCase_ : int): """simple docstring""" if self.label_column not in features: raise ValueError(F"Column {self.label_column} is not present in features.") if not isinstance(features[self.label_column] , UpperCamelCase_): raise ValueError(F"Column {self.label_column} is not a ClassLabel.") __UpperCAmelCase : int = copy.deepcopy(self) __UpperCAmelCase : Optional[int] = self.label_schema.copy() __UpperCAmelCase : Any = features[self.label_column] __UpperCAmelCase : List[str] = label_schema return task_template @property def a_ ( self : Any): """simple docstring""" return { self.text_column: "text", self.label_column: "labels", }
77
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore A = """ Human: <<task>> Assistant: """ A = """huggingface-tools/default-prompts""" A = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase="run" ) -> List[str]: """simple docstring""" if prompt_or_repo_id is None: __UpperCAmelCase : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , UpperCamelCase ) is not None: return prompt_or_repo_id __UpperCAmelCase : str = cached_file( UpperCamelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(UpperCamelCase , "r" , encoding="utf-8" ) as f: return f.read()
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { """microsoft/trocr-base-handwritten""": ( """https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json""" ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class a__ ( __magic_name__ ): lowercase_ = "trocr" lowercase_ = ["past_key_values"] lowercase_ = { "num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model", "num_hidden_layers": "decoder_layers", } def __init__( self : Optional[Any] , UpperCamelCase_ : List[Any]=50265 , UpperCamelCase_ : Any=1024 , UpperCamelCase_ : Optional[Any]=12 , UpperCamelCase_ : Any=16 , UpperCamelCase_ : Optional[Any]=4096 , UpperCamelCase_ : str="gelu" , UpperCamelCase_ : List[str]=512 , UpperCamelCase_ : List[str]=0.1 , UpperCamelCase_ : Optional[Any]=0.0 , UpperCamelCase_ : str=0.0 , UpperCamelCase_ : str=2 , UpperCamelCase_ : List[str]=0.02 , UpperCamelCase_ : Tuple=0.0 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Dict=False , UpperCamelCase_ : Any=True , UpperCamelCase_ : Dict=True , UpperCamelCase_ : str=1 , UpperCamelCase_ : str=0 , UpperCamelCase_ : List[str]=2 , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" __UpperCAmelCase : List[Any] = vocab_size __UpperCAmelCase : str = d_model __UpperCAmelCase : Tuple = decoder_layers __UpperCAmelCase : Optional[Any] = decoder_attention_heads __UpperCAmelCase : Optional[int] = decoder_ffn_dim __UpperCAmelCase : Optional[int] = activation_function __UpperCAmelCase : Optional[int] = max_position_embeddings __UpperCAmelCase : Union[str, Any] = dropout __UpperCAmelCase : str = attention_dropout __UpperCAmelCase : Any = activation_dropout __UpperCAmelCase : Any = init_std __UpperCAmelCase : List[str] = decoder_layerdrop __UpperCAmelCase : Tuple = use_cache __UpperCAmelCase : str = scale_embedding __UpperCAmelCase : str = use_learned_position_embeddings __UpperCAmelCase : List[Any] = layernorm_embedding super().__init__( pad_token_id=UpperCamelCase_ , bos_token_id=UpperCamelCase_ , eos_token_id=UpperCamelCase_ , decoder_start_token_id=UpperCamelCase_ , **UpperCamelCase_ , )
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available A = { """configuration_ernie""": ["""ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ErnieConfig""", """ErnieOnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST""", """ErnieForCausalLM""", """ErnieForMaskedLM""", """ErnieForMultipleChoice""", """ErnieForNextSentencePrediction""", """ErnieForPreTraining""", """ErnieForQuestionAnswering""", """ErnieForSequenceClassification""", """ErnieForTokenClassification""", """ErnieModel""", """ErniePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { """sayakpaul/vit-msn-base""": """https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json""", # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class a__ ( __magic_name__ ): lowercase_ = "vit_msn" def __init__( self : int , UpperCamelCase_ : Optional[int]=768 , UpperCamelCase_ : str=12 , UpperCamelCase_ : Tuple=12 , UpperCamelCase_ : int=3072 , UpperCamelCase_ : Any="gelu" , UpperCamelCase_ : Optional[int]=0.0 , UpperCamelCase_ : Tuple=0.0 , UpperCamelCase_ : Any=0.02 , UpperCamelCase_ : Tuple=1e-06 , UpperCamelCase_ : Tuple=224 , UpperCamelCase_ : Tuple=16 , UpperCamelCase_ : Any=3 , UpperCamelCase_ : Any=True , **UpperCamelCase_ : Tuple , ): """simple docstring""" super().__init__(**UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = hidden_size __UpperCAmelCase : str = num_hidden_layers __UpperCAmelCase : List[str] = num_attention_heads __UpperCAmelCase : int = intermediate_size __UpperCAmelCase : List[Any] = hidden_act __UpperCAmelCase : Union[str, Any] = hidden_dropout_prob __UpperCAmelCase : Optional[int] = attention_probs_dropout_prob __UpperCAmelCase : int = initializer_range __UpperCAmelCase : Dict = layer_norm_eps __UpperCAmelCase : Optional[int] = image_size __UpperCAmelCase : Optional[Any] = patch_size __UpperCAmelCase : Any = num_channels __UpperCAmelCase : List[Any] = qkv_bias
77
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class a__ ( nn.Module ): def __init__( self : Union[str, Any]): """simple docstring""" super().__init__() __UpperCAmelCase : Optional[int] = nn.Linear(3 , 4) __UpperCAmelCase : str = nn.BatchNormad(4) __UpperCAmelCase : int = nn.Linear(4 , 5) def a_ ( self : str , UpperCamelCase_ : List[str]): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCamelCase_))) class a__ ( unittest.TestCase ): def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Optional[Any] = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , model.state_dict()) __UpperCAmelCase : Union[str, Any] = os.path.join(UpperCamelCase_ , "index.json") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: __UpperCAmelCase : Optional[int] = os.path.join(UpperCamelCase_ , F"{key}.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on the fact weights are properly loaded def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : int = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: __UpperCAmelCase : List[Any] = torch.randn(2 , 3 , dtype=UpperCamelCase_) with TemporaryDirectory() as tmp_dir: __UpperCAmelCase : Tuple = offload_weight(UpperCamelCase_ , "weight" , UpperCamelCase_ , {}) __UpperCAmelCase : Dict = os.path.join(UpperCamelCase_ , "weight.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) self.assertDictEqual(UpperCamelCase_ , {"weight": {"shape": [2, 3], "dtype": str(UpperCamelCase_).split(".")[1]}}) __UpperCAmelCase : Optional[Any] = load_offloaded_weight(UpperCamelCase_ , index["weight"]) self.assertTrue(torch.equal(UpperCamelCase_ , UpperCamelCase_)) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : List[Any] = ModelForTest() __UpperCAmelCase : Optional[int] = model.state_dict() __UpperCAmelCase : List[str] = {k: v for k, v in state_dict.items() if "linear2" not in k} __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "linear2" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : List[str] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "weight" in k} __UpperCAmelCase : Optional[Any] = {k: v for k, v in state_dict.items() if "weight" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[Any] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) # Duplicates are removed __UpperCAmelCase : str = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Any = {"a.1": 0, "a.10": 1, "a.2": 2} __UpperCAmelCase : Union[str, Any] = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1": 0, "a.2": 2}) __UpperCAmelCase : int = {"a.1.a": 0, "a.10.a": 1, "a.2.a": 2} __UpperCAmelCase : int = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1.a": 0, "a.2.a": 2})
77
1
"""simple docstring""" from collections import deque from .hash_table import HashTable class a__ ( __magic_name__ ): def __init__( self : List[Any] , *UpperCamelCase_ : Union[str, Any] , **UpperCamelCase_ : List[str]): """simple docstring""" super().__init__(*UpperCamelCase_ , **UpperCamelCase_) def a_ ( self : str , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : List[str] = deque([]) if self.values[key] is None else self.values[key] self.values[key].appendleft(UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = self.values[key] def a_ ( self : List[str]): """simple docstring""" return ( sum(self.charge_factor - len(UpperCamelCase_) for slot in self.values) / self.size_table * self.charge_factor ) def a_ ( self : str , UpperCamelCase_ : List[str] , UpperCamelCase_ : Optional[int]=None): """simple docstring""" if not ( len(self.values[key]) == self.charge_factor and self.values.count(UpperCamelCase_) == 0 ): return key return super()._collision_resolution(UpperCamelCase_ , UpperCamelCase_)
77
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Dict = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __UpperCAmelCase : Union[str, Any] = n - k # Calculate C(n,k) for i in range(UpperCamelCase ): result *= n - i result //= i + 1 return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , UpperCamelCase ) // (node_count + 1) def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if n < 0: raise ValueError("factorial() not defined for negative values" ) __UpperCAmelCase : Optional[Any] = 1 for i in range(1 , n + 1 ): result *= i return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return catalan_number(UpperCamelCase ) * factorial(UpperCamelCase ) if __name__ == "__main__": A = int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
77
1
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_features", "is_longer"] def __init__( self : List[str] , UpperCamelCase_ : Dict=64 , UpperCamelCase_ : Tuple=48000 , UpperCamelCase_ : List[Any]=480 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=1024 , UpperCamelCase_ : List[str]=0.0 , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : float = 0 , UpperCamelCase_ : float = 14000 , UpperCamelCase_ : int = None , UpperCamelCase_ : str = "fusion" , UpperCamelCase_ : str = "repeatpad" , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" super().__init__( feature_size=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , padding_value=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = top_db __UpperCAmelCase : Optional[Any] = truncation __UpperCAmelCase : str = padding __UpperCAmelCase : int = fft_window_size __UpperCAmelCase : str = (fft_window_size >> 1) + 1 __UpperCAmelCase : List[Any] = hop_length __UpperCAmelCase : Optional[Any] = max_length_s __UpperCAmelCase : Tuple = max_length_s * sampling_rate __UpperCAmelCase : str = sampling_rate __UpperCAmelCase : int = frequency_min __UpperCAmelCase : Optional[Any] = frequency_max __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm=UpperCamelCase_ , mel_scale="htk" , ) __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm="slaney" , mel_scale="slaney" , ) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Dict = copy.deepcopy(self.__dict__) __UpperCAmelCase : str = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def a_ ( self : int , UpperCamelCase_ : np.array , UpperCamelCase_ : Optional[np.array] = None): """simple docstring""" __UpperCAmelCase : List[Any] = spectrogram( UpperCamelCase_ , window_function(self.fft_window_size , "hann") , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=UpperCamelCase_ , log_mel="dB" , ) return log_mel_spectrogram.T def a_ ( self : List[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = np.array_split(list(range(0 , total_frames - chunk_frames + 1)) , 3) if len(ranges[1]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : str = [0] if len(ranges[2]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : Dict = [0] # randomly choose index for each part __UpperCAmelCase : Dict = np.random.choice(ranges[0]) __UpperCAmelCase : List[str] = np.random.choice(ranges[1]) __UpperCAmelCase : List[Any] = np.random.choice(ranges[2]) __UpperCAmelCase : List[Any] = mel[idx_front : idx_front + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_back : idx_back + chunk_frames, :] __UpperCAmelCase : Tuple = torch.tensor(mel[None, None, :]) __UpperCAmelCase : Union[str, Any] = torch.nn.functional.interpolate( UpperCamelCase_ , size=[chunk_frames, 64] , mode="bilinear" , align_corners=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = mel_shrink[0][0].numpy() __UpperCAmelCase : Optional[int] = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0) return mel_fusion def a_ ( self : Optional[Any] , UpperCamelCase_ : np.array , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[Any]): """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": __UpperCAmelCase : List[str] = True # random crop to max_length (for compatibility) -> this should be handled by self.pad __UpperCAmelCase : List[Any] = len(UpperCamelCase_) - max_length __UpperCAmelCase : int = np.random.randint(0 , overflow + 1) __UpperCAmelCase : Union[str, Any] = waveform[idx : idx + max_length] __UpperCAmelCase : Union[str, Any] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] elif truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : Dict = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed __UpperCAmelCase : Tuple = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. __UpperCAmelCase : List[str] = np.stack([mel, mel, mel, mel] , axis=0) __UpperCAmelCase : Any = False else: __UpperCAmelCase : List[str] = self._random_mel_fusion(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented") else: __UpperCAmelCase : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": __UpperCAmelCase : Tuple = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : List[str] = np.stack(np.tile(UpperCamelCase_ , n_repeat + 1))[:max_length] if padding == "repeatpad": __UpperCAmelCase : Union[str, Any] = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : Optional[Any] = np.stack(np.tile(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : int = np.pad(UpperCamelCase_ , (0, max_length - waveform.shape[0]) , mode="constant" , constant_values=0) if truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0) else: __UpperCAmelCase : Optional[int] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] return input_mel, longer def __call__( self : Dict , UpperCamelCase_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase_ : str = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[Union[str, TensorType]] = None , **UpperCamelCase_ : Any , ): """simple docstring""" __UpperCAmelCase : int = truncation if truncation is not None else self.truncation __UpperCAmelCase : Optional[Any] = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}.") else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug.") __UpperCAmelCase : List[str] = isinstance(UpperCamelCase_ , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}") __UpperCAmelCase : str = is_batched_numpy or ( isinstance(UpperCamelCase_ , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: __UpperCAmelCase : Dict = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase_ , np.ndarray): __UpperCAmelCase : Tuple = np.asarray(UpperCamelCase_ , dtype=np.floataa) elif isinstance(UpperCamelCase_ , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): __UpperCAmelCase : Optional[int] = raw_speech.astype(np.floataa) # always return batch if not is_batched: __UpperCAmelCase : int = [np.asarray(UpperCamelCase_)] # convert to mel spectrogram, truncate and pad if needed. __UpperCAmelCase : Optional[int] = [ self._get_input_mel(UpperCamelCase_ , max_length if max_length else self.nb_max_samples , UpperCamelCase_ , UpperCamelCase_) for waveform in raw_speech ] __UpperCAmelCase : Tuple = [] __UpperCAmelCase : List[Any] = [] for mel, longer in padded_inputs: input_mel.append(UpperCamelCase_) is_longer.append(UpperCamelCase_) if truncation == "fusion" and sum(UpperCamelCase_) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer __UpperCAmelCase : Any = np.random.randint(0 , len(UpperCamelCase_)) __UpperCAmelCase : Optional[int] = True if isinstance(input_mel[0] , UpperCamelCase_): __UpperCAmelCase : Tuple = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for feature in input_mel] # is_longer is a list of bool __UpperCAmelCase : List[str] = [[longer] for longer in is_longer] __UpperCAmelCase : Optional[int] = {"input_features": input_mel, "is_longer": is_longer} __UpperCAmelCase : Optional[int] = BatchFeature(UpperCamelCase_) if return_tensors is not None: __UpperCAmelCase : Any = input_features.convert_to_tensors(UpperCamelCase_) return input_features
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) A = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" from abc import ABC, abstractmethod from argparse import ArgumentParser class a__ ( __magic_name__ ): @staticmethod @abstractmethod def a_ ( UpperCamelCase_ : ArgumentParser): """simple docstring""" raise NotImplementedError() @abstractmethod def a_ ( self : List[Any]): """simple docstring""" raise NotImplementedError()
77
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCamelCase_ : int , UpperCamelCase_ : Any=13 , UpperCamelCase_ : Optional[int]=3 , UpperCamelCase_ : int=224 , UpperCamelCase_ : int=30 , UpperCamelCase_ : str=400 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Optional[int]=[0.5, 0.5, 0.5] , UpperCamelCase_ : Optional[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __UpperCAmelCase : Tuple = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : List[Any] = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : List[Any] = image_size __UpperCAmelCase : str = min_resolution __UpperCAmelCase : Tuple = max_resolution __UpperCAmelCase : Optional[Any] = do_resize __UpperCAmelCase : Any = size __UpperCAmelCase : Any = do_normalize __UpperCAmelCase : Any = image_mean __UpperCAmelCase : Optional[Any] = image_std def a_ ( self : str): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ViTImageProcessor if is_vision_available() else None def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Optional[Any] = EfficientFormerImageProcessorTester(self) @property def a_ ( self : Union[str, Any]): """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCamelCase_ , "image_std")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) def a_ ( self : Dict): """simple docstring""" pass def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : str = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Optional[int] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : Tuple = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Any = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
77
1
"""simple docstring""" import heapq def _UpperCamelCase ( UpperCamelCase ) -> set[int]: """simple docstring""" __UpperCAmelCase : list[list] = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(UpperCamelCase , [-1 * len(UpperCamelCase ), (key, value)] ) # chosen_vertices = set of chosen vertices __UpperCAmelCase : List[str] = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices __UpperCAmelCase : Tuple = heapq.heappop(UpperCamelCase )[1][0] chosen_vertices.add(UpperCamelCase ) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: __UpperCAmelCase : str = elem[1][1].index(UpperCamelCase ) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(UpperCamelCase ) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() A = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(f'''Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}''')
77
"""simple docstring""" from collections import namedtuple A = namedtuple("""from_to""", """from_ to""") A = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1_000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": from_to(0.76455, 1.30795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.000236588, 4226.75), } def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'from_type' value: {from_type!r} Supported values are:\n" + ", ".join(UpperCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n" + ", ".join(UpperCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
77
1
"""simple docstring""" from __future__ import annotations from cmath import sqrt def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> tuple[complex, complex]: """simple docstring""" if a == 0: raise ValueError("Coefficient 'a' must not be zero." ) __UpperCAmelCase : str = b * b - 4 * a * c __UpperCAmelCase : Any = (-b + sqrt(UpperCamelCase )) / (2 * a) __UpperCAmelCase : Union[str, Any] = (-b - sqrt(UpperCamelCase )) / (2 * a) return ( root_a.real if not root_a.imag else root_a, root_a.real if not root_a.imag else root_a, ) def _UpperCamelCase ( ) -> Tuple: """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : List[str] = quadratic_roots(a=5 , b=6 , c=1 ) print(f"The solutions are: {solutiona} and {solutiona}" ) if __name__ == "__main__": main()
77
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ShapEPipeline lowercase_ = ["prompt"] lowercase_ = ["prompt"] lowercase_ = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] lowercase_ = False @property def a_ ( self : Optional[int]): """simple docstring""" return 32 @property def a_ ( self : Any): """simple docstring""" return 32 @property def a_ ( self : int): """simple docstring""" return self.time_input_dim * 4 @property def a_ ( self : List[Any]): """simple docstring""" return 8 @property def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip") return tokenizer @property def a_ ( self : List[str]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModelWithProjection(UpperCamelCase_) @property def a_ ( self : Any): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Union[str, Any] = { "num_attention_heads": 2, "attention_head_dim": 16, "embedding_dim": self.time_input_dim, "num_embeddings": 32, "embedding_proj_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "num_layers": 1, "clip_embed_dim": self.time_input_dim * 2, "additional_embeddings": 0, "time_embed_act_fn": "gelu", "norm_in_type": "layer", "encoder_hid_proj_type": None, "added_emb_type": None, } __UpperCAmelCase : Dict = PriorTransformer(**UpperCamelCase_) return model @property def a_ ( self : Union[str, Any]): """simple docstring""" torch.manual_seed(0) __UpperCAmelCase : Tuple = { "param_shapes": ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), "d_latent": self.time_input_dim, "d_hidden": self.renderer_dim, "n_output": 12, "background": ( 0.1, 0.1, 0.1, ), } __UpperCAmelCase : List[Any] = ShapERenderer(**UpperCamelCase_) return model def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.dummy_prior __UpperCAmelCase : str = self.dummy_text_encoder __UpperCAmelCase : int = self.dummy_tokenizer __UpperCAmelCase : int = self.dummy_renderer __UpperCAmelCase : Tuple = HeunDiscreteScheduler( beta_schedule="exp" , num_train_timesteps=1024 , prediction_type="sample" , use_karras_sigmas=UpperCamelCase_ , clip_sample=UpperCamelCase_ , clip_sample_range=1.0 , ) __UpperCAmelCase : str = { "prior": prior, "text_encoder": text_encoder, "tokenizer": tokenizer, "renderer": renderer, "scheduler": scheduler, } return components def a_ ( self : Optional[int] , UpperCamelCase_ : Dict , UpperCamelCase_ : Any=0): """simple docstring""" if str(UpperCamelCase_).startswith("mps"): __UpperCAmelCase : List[Any] = torch.manual_seed(UpperCamelCase_) else: __UpperCAmelCase : str = torch.Generator(device=UpperCamelCase_).manual_seed(UpperCamelCase_) __UpperCAmelCase : List[Any] = { "prompt": "horse", "generator": generator, "num_inference_steps": 1, "frame_size": 32, "output_type": "np", } return inputs def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : str = "cpu" __UpperCAmelCase : Union[str, Any] = self.get_dummy_components() __UpperCAmelCase : Union[str, Any] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[Any] = pipe(**self.get_dummy_inputs(UpperCamelCase_)) __UpperCAmelCase : Union[str, Any] = output.images[0] __UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) __UpperCAmelCase : Union[str, Any] = np.array( [ 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, 0.00039216, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def a_ ( self : Tuple): """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2]) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch_device == "cpu" __UpperCAmelCase : Optional[Any] = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=UpperCamelCase_ , relax_max_difference=UpperCamelCase_ , ) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.get_dummy_components() __UpperCAmelCase : List[str] = self.pipeline_class(**UpperCamelCase_) __UpperCAmelCase : int = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Optional[int] = 1 __UpperCAmelCase : Any = 2 __UpperCAmelCase : Optional[Any] = self.get_dummy_inputs(UpperCamelCase_) for key in inputs.keys(): if key in self.batch_params: __UpperCAmelCase : List[Any] = batch_size * [inputs[key]] __UpperCAmelCase : List[Any] = pipe(**UpperCamelCase_ , num_images_per_prompt=UpperCamelCase_)[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class a__ ( unittest.TestCase ): def a_ ( self : List[str]): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/shap_e/test_shap_e_np_out.npy") __UpperCAmelCase : Optional[Any] = ShapEPipeline.from_pretrained("openai/shap-e") __UpperCAmelCase : Any = pipe.to(UpperCamelCase_) pipe.set_progress_bar_config(disable=UpperCamelCase_) __UpperCAmelCase : Dict = torch.Generator(device=UpperCamelCase_).manual_seed(0) __UpperCAmelCase : int = pipe( "a shark" , generator=UpperCamelCase_ , guidance_scale=15.0 , num_inference_steps=64 , frame_size=64 , output_type="np" , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(UpperCamelCase_ , UpperCamelCase_)
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("iterations must be defined as integers" ) if not isinstance(UpperCamelCase , UpperCamelCase ) or not number >= 1: raise ValueError( "starting number must be\n and integer and be more than 0" ) if not iterations >= 1: raise ValueError("Iterations must be done more than 0 times to play FizzBuzz" ) __UpperCAmelCase : Tuple = "" while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(UpperCamelCase ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_features", "is_longer"] def __init__( self : List[str] , UpperCamelCase_ : Dict=64 , UpperCamelCase_ : Tuple=48000 , UpperCamelCase_ : List[Any]=480 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=1024 , UpperCamelCase_ : List[str]=0.0 , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : float = 0 , UpperCamelCase_ : float = 14000 , UpperCamelCase_ : int = None , UpperCamelCase_ : str = "fusion" , UpperCamelCase_ : str = "repeatpad" , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" super().__init__( feature_size=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , padding_value=UpperCamelCase_ , return_attention_mask=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Union[str, Any] = top_db __UpperCAmelCase : Optional[Any] = truncation __UpperCAmelCase : str = padding __UpperCAmelCase : int = fft_window_size __UpperCAmelCase : str = (fft_window_size >> 1) + 1 __UpperCAmelCase : List[Any] = hop_length __UpperCAmelCase : Optional[Any] = max_length_s __UpperCAmelCase : Tuple = max_length_s * sampling_rate __UpperCAmelCase : str = sampling_rate __UpperCAmelCase : int = frequency_min __UpperCAmelCase : Optional[Any] = frequency_max __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm=UpperCamelCase_ , mel_scale="htk" , ) __UpperCAmelCase : Any = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=UpperCamelCase_ , min_frequency=UpperCamelCase_ , max_frequency=UpperCamelCase_ , sampling_rate=UpperCamelCase_ , norm="slaney" , mel_scale="slaney" , ) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Dict = copy.deepcopy(self.__dict__) __UpperCAmelCase : str = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def a_ ( self : int , UpperCamelCase_ : np.array , UpperCamelCase_ : Optional[np.array] = None): """simple docstring""" __UpperCAmelCase : List[Any] = spectrogram( UpperCamelCase_ , window_function(self.fft_window_size , "hann") , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=UpperCamelCase_ , log_mel="dB" , ) return log_mel_spectrogram.T def a_ ( self : List[Any] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = np.array_split(list(range(0 , total_frames - chunk_frames + 1)) , 3) if len(ranges[1]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : str = [0] if len(ranges[2]) == 0: # if the audio is too short, we just use the first chunk __UpperCAmelCase : Dict = [0] # randomly choose index for each part __UpperCAmelCase : Dict = np.random.choice(ranges[0]) __UpperCAmelCase : List[str] = np.random.choice(ranges[1]) __UpperCAmelCase : List[Any] = np.random.choice(ranges[2]) __UpperCAmelCase : List[Any] = mel[idx_front : idx_front + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_middle : idx_middle + chunk_frames, :] __UpperCAmelCase : List[str] = mel[idx_back : idx_back + chunk_frames, :] __UpperCAmelCase : Tuple = torch.tensor(mel[None, None, :]) __UpperCAmelCase : Union[str, Any] = torch.nn.functional.interpolate( UpperCamelCase_ , size=[chunk_frames, 64] , mode="bilinear" , align_corners=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = mel_shrink[0][0].numpy() __UpperCAmelCase : Optional[int] = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0) return mel_fusion def a_ ( self : Optional[Any] , UpperCamelCase_ : np.array , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[Any]): """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": __UpperCAmelCase : List[str] = True # random crop to max_length (for compatibility) -> this should be handled by self.pad __UpperCAmelCase : List[Any] = len(UpperCamelCase_) - max_length __UpperCAmelCase : int = np.random.randint(0 , overflow + 1) __UpperCAmelCase : Union[str, Any] = waveform[idx : idx + max_length] __UpperCAmelCase : Union[str, Any] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] elif truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : Dict = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed __UpperCAmelCase : Tuple = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. __UpperCAmelCase : List[str] = np.stack([mel, mel, mel, mel] , axis=0) __UpperCAmelCase : Any = False else: __UpperCAmelCase : List[str] = self._random_mel_fusion(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented") else: __UpperCAmelCase : Optional[Any] = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": __UpperCAmelCase : Tuple = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : List[str] = np.stack(np.tile(UpperCamelCase_ , n_repeat + 1))[:max_length] if padding == "repeatpad": __UpperCAmelCase : Union[str, Any] = int(max_length / len(UpperCamelCase_)) __UpperCAmelCase : Optional[Any] = np.stack(np.tile(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : int = np.pad(UpperCamelCase_ , (0, max_length - waveform.shape[0]) , mode="constant" , constant_values=0) if truncation == "fusion": __UpperCAmelCase : Any = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters) __UpperCAmelCase : List[Any] = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0) else: __UpperCAmelCase : Optional[int] = self._np_extract_fbank_features(UpperCamelCase_ , self.mel_filters_slaney)[None, :] return input_mel, longer def __call__( self : Dict , UpperCamelCase_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase_ : str = None , UpperCamelCase_ : Optional[str] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[int] = None , UpperCamelCase_ : Optional[Union[str, TensorType]] = None , **UpperCamelCase_ : Any , ): """simple docstring""" __UpperCAmelCase : int = truncation if truncation is not None else self.truncation __UpperCAmelCase : Optional[Any] = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}.") else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug.") __UpperCAmelCase : List[str] = isinstance(UpperCamelCase_ , np.ndarray) and len(raw_speech.shape) > 1 if is_batched_numpy and len(raw_speech.shape) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}") __UpperCAmelCase : str = is_batched_numpy or ( isinstance(UpperCamelCase_ , (list, tuple)) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list))) ) if is_batched: __UpperCAmelCase : Dict = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase_ , np.ndarray): __UpperCAmelCase : Tuple = np.asarray(UpperCamelCase_ , dtype=np.floataa) elif isinstance(UpperCamelCase_ , np.ndarray) and raw_speech.dtype is np.dtype(np.floataa): __UpperCAmelCase : Optional[int] = raw_speech.astype(np.floataa) # always return batch if not is_batched: __UpperCAmelCase : int = [np.asarray(UpperCamelCase_)] # convert to mel spectrogram, truncate and pad if needed. __UpperCAmelCase : Optional[int] = [ self._get_input_mel(UpperCamelCase_ , max_length if max_length else self.nb_max_samples , UpperCamelCase_ , UpperCamelCase_) for waveform in raw_speech ] __UpperCAmelCase : Tuple = [] __UpperCAmelCase : List[Any] = [] for mel, longer in padded_inputs: input_mel.append(UpperCamelCase_) is_longer.append(UpperCamelCase_) if truncation == "fusion" and sum(UpperCamelCase_) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer __UpperCAmelCase : Any = np.random.randint(0 , len(UpperCamelCase_)) __UpperCAmelCase : Optional[int] = True if isinstance(input_mel[0] , UpperCamelCase_): __UpperCAmelCase : Tuple = [np.asarray(UpperCamelCase_ , dtype=np.floataa) for feature in input_mel] # is_longer is a list of bool __UpperCAmelCase : List[str] = [[longer] for longer in is_longer] __UpperCAmelCase : Optional[int] = {"input_features": input_mel, "is_longer": is_longer} __UpperCAmelCase : Optional[int] = BatchFeature(UpperCamelCase_) if return_tensors is not None: __UpperCAmelCase : Any = input_features.convert_to_tensors(UpperCamelCase_) return input_features
77
1
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A = logging.get_logger(__name__) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=False ) -> Tuple: """simple docstring""" __UpperCAmelCase : Tuple = OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith("head" ): __UpperCAmelCase : str = "segformer.encoder." + key if key.startswith("backbone" ): __UpperCAmelCase : Union[str, Any] = key.replace("backbone" , "segformer.encoder" ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 __UpperCAmelCase : str = key[key.find("patch_embed" ) + len("patch_embed" )] __UpperCAmelCase : Any = key.replace(f"patch_embed{idx}" , f"patch_embeddings.{int(UpperCamelCase )-1}" ) if "norm" in key: __UpperCAmelCase : List[str] = key.replace("norm" , "layer_norm" ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 __UpperCAmelCase : int = key[key.find("segformer.encoder.layer_norm" ) + len("segformer.encoder.layer_norm" )] __UpperCAmelCase : Optional[Any] = key.replace(f"layer_norm{idx}" , f"layer_norm.{int(UpperCamelCase )-1}" ) if "layer_norm1" in key: __UpperCAmelCase : Union[str, Any] = key.replace("layer_norm1" , "layer_norm_1" ) if "layer_norm2" in key: __UpperCAmelCase : Optional[int] = key.replace("layer_norm2" , "layer_norm_2" ) if "block" in key: # replace for example block1 by block.0 __UpperCAmelCase : Union[str, Any] = key[key.find("block" ) + len("block" )] __UpperCAmelCase : List[str] = key.replace(f"block{idx}" , f"block.{int(UpperCamelCase )-1}" ) if "attn.q" in key: __UpperCAmelCase : List[str] = key.replace("attn.q" , "attention.self.query" ) if "attn.proj" in key: __UpperCAmelCase : List[Any] = key.replace("attn.proj" , "attention.output.dense" ) if "attn" in key: __UpperCAmelCase : Union[str, Any] = key.replace("attn" , "attention.self" ) if "fc1" in key: __UpperCAmelCase : int = key.replace("fc1" , "dense1" ) if "fc2" in key: __UpperCAmelCase : str = key.replace("fc2" , "dense2" ) if "linear_pred" in key: __UpperCAmelCase : Optional[int] = key.replace("linear_pred" , "classifier" ) if "linear_fuse" in key: __UpperCAmelCase : Tuple = key.replace("linear_fuse.conv" , "linear_fuse" ) __UpperCAmelCase : Tuple = key.replace("linear_fuse.bn" , "batch_norm" ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 __UpperCAmelCase : Tuple = key[key.find("linear_c" ) + len("linear_c" )] __UpperCAmelCase : Any = key.replace(f"linear_c{idx}" , f"linear_c.{int(UpperCamelCase )-1}" ) if key.startswith("head" ): __UpperCAmelCase : Optional[int] = key.replace("head" , "classifier" ) __UpperCAmelCase : List[Any] = value return new_state_dict def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" # for each of the encoder blocks: for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) __UpperCAmelCase : Tuple = state_dict.pop(f"segformer.encoder.block.{i}.{j}.attention.self.kv.weight" ) __UpperCAmelCase : Optional[int] = state_dict.pop(f"segformer.encoder.block.{i}.{j}.attention.self.kv.bias" ) # next, add keys and values (in that order) to the state dict __UpperCAmelCase : str = kv_weight[ : config.hidden_sizes[i], : ] __UpperCAmelCase : List[str] = kv_bias[: config.hidden_sizes[i]] __UpperCAmelCase : int = kv_weight[ config.hidden_sizes[i] :, : ] __UpperCAmelCase : int = kv_bias[ config.hidden_sizes[i] : ] def _UpperCamelCase ( ) -> Optional[Any]: """simple docstring""" __UpperCAmelCase : List[Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" __UpperCAmelCase : Optional[Any] = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ) return image @torch.no_grad() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : Optional[Any] = SegformerConfig() __UpperCAmelCase : Optional[int] = False # set attributes based on model_name __UpperCAmelCase : Dict = "huggingface/label-files" if "segformer" in model_name: __UpperCAmelCase : Any = model_name[len("segformer." ) : len("segformer." ) + 2] if "ade" in model_name: __UpperCAmelCase : Dict = 150 __UpperCAmelCase : Dict = "ade20k-id2label.json" __UpperCAmelCase : List[Any] = (1, 150, 128, 128) elif "city" in model_name: __UpperCAmelCase : Optional[int] = 19 __UpperCAmelCase : Optional[Any] = "cityscapes-id2label.json" __UpperCAmelCase : Tuple = (1, 19, 128, 128) else: raise ValueError(f"Model {model_name} not supported" ) elif "mit" in model_name: __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : str = model_name[4:6] __UpperCAmelCase : Tuple = 1000 __UpperCAmelCase : Any = "imagenet-1k-id2label.json" __UpperCAmelCase : List[str] = (1, 1000) else: raise ValueError(f"Model {model_name} not supported" ) # set config attributes __UpperCAmelCase : List[Any] = json.load(open(hf_hub_download(UpperCamelCase , UpperCamelCase , repo_type="dataset" ) , "r" ) ) __UpperCAmelCase : Optional[Any] = {int(UpperCamelCase ): v for k, v in idalabel.items()} __UpperCAmelCase : List[Any] = idalabel __UpperCAmelCase : List[str] = {v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": __UpperCAmelCase : Optional[int] = [64, 128, 320, 512] __UpperCAmelCase : Tuple = 256 elif size == "b2": __UpperCAmelCase : int = [64, 128, 320, 512] __UpperCAmelCase : List[str] = 768 __UpperCAmelCase : Optional[int] = [3, 4, 6, 3] elif size == "b3": __UpperCAmelCase : Tuple = [64, 128, 320, 512] __UpperCAmelCase : List[Any] = 768 __UpperCAmelCase : List[str] = [3, 4, 18, 3] elif size == "b4": __UpperCAmelCase : Any = [64, 128, 320, 512] __UpperCAmelCase : Dict = 768 __UpperCAmelCase : List[str] = [3, 8, 27, 3] elif size == "b5": __UpperCAmelCase : str = [64, 128, 320, 512] __UpperCAmelCase : List[Any] = 768 __UpperCAmelCase : List[str] = [3, 6, 40, 3] else: raise ValueError(f"Size {size} not supported" ) # load image processor (only resize + normalize) __UpperCAmelCase : str = SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=UpperCamelCase , align=UpperCamelCase , do_random_crop=UpperCamelCase ) # prepare image __UpperCAmelCase : str = prepare_img() __UpperCAmelCase : str = image_processor(images=UpperCamelCase , return_tensors="pt" ).pixel_values logger.info(f"Converting model {model_name}..." ) # load original state dict if encoder_only: __UpperCAmelCase : int = torch.load(UpperCamelCase , map_location=torch.device("cpu" ) ) else: __UpperCAmelCase : str = torch.load(UpperCamelCase , map_location=torch.device("cpu" ) )["state_dict"] # rename keys __UpperCAmelCase : int = rename_keys(UpperCamelCase , encoder_only=UpperCamelCase ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(UpperCamelCase , UpperCamelCase ) # create HuggingFace model and load state dict if encoder_only: __UpperCAmelCase : int = False __UpperCAmelCase : Optional[Any] = SegformerForImageClassification(UpperCamelCase ) else: __UpperCAmelCase : Tuple = SegformerForSemanticSegmentation(UpperCamelCase ) model.load_state_dict(UpperCamelCase ) model.eval() # forward pass __UpperCAmelCase : str = model(UpperCamelCase ) __UpperCAmelCase : Any = outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": __UpperCAmelCase : List[str] = torch.tensor( [ [[-4.6310, -5.5232, -6.2356], [-5.1921, -6.1444, -6.5996], [-5.4424, -6.2790, -6.7574]], [[-12.1391, -13.3122, -13.9554], [-12.8732, -13.9352, -14.3563], [-12.9438, -13.8226, -14.2513]], [[-12.5134, -13.4686, -14.4915], [-12.8669, -14.4343, -14.7758], [-13.2523, -14.5819, -15.0694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": __UpperCAmelCase : Any = torch.tensor( [ [[-7.5820, -8.7231, -8.3215], [-8.0600, -10.3529, -10.0304], [-7.5208, -9.4103, -9.6239]], [[-12.6918, -13.8994, -13.7137], [-13.3196, -15.7523, -15.4789], [-12.9343, -14.8757, -14.9689]], [[-11.1911, -11.9421, -11.3243], [-11.3342, -13.6839, -13.3581], [-10.3909, -12.1832, -12.4858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": __UpperCAmelCase : Dict = torch.tensor( [ [[-11.8173, -14.3850, -16.3128], [-14.5648, -16.5804, -18.6568], [-14.7223, -15.7387, -18.4218]], [[-15.7290, -17.9171, -19.4423], [-18.3105, -19.9448, -21.4661], [-17.9296, -18.6497, -20.7910]], [[-15.0783, -17.0336, -18.2789], [-16.8771, -18.6870, -20.1612], [-16.2454, -17.1426, -19.5055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": __UpperCAmelCase : str = torch.tensor( [ [[-9.0878, -10.2081, -10.1891], [-9.3144, -10.7941, -10.9843], [-9.2294, -10.3855, -10.5704]], [[-12.2316, -13.9068, -13.6102], [-12.9161, -14.3702, -14.3235], [-12.5233, -13.7174, -13.7932]], [[-14.6275, -15.2490, -14.9727], [-14.3400, -15.9687, -16.2827], [-14.1484, -15.4033, -15.8937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": __UpperCAmelCase : List[Any] = torch.tensor( [ [[-12.3144, -13.2447, -14.0802], [-13.3614, -14.5816, -15.6117], [-13.3340, -14.4433, -16.2219]], [[-19.2781, -20.4128, -20.7506], [-20.6153, -21.6566, -22.0998], [-19.9800, -21.0430, -22.1494]], [[-18.8739, -19.7804, -21.1834], [-20.1233, -21.6765, -23.2944], [-20.0315, -21.2641, -23.6944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": __UpperCAmelCase : Dict = torch.tensor( [ [[-9.5524, -12.0835, -11.7348], [-10.5229, -13.6446, -14.5662], [-9.5842, -12.8851, -13.9414]], [[-15.3432, -17.5323, -17.0818], [-16.3330, -18.9255, -19.2101], [-15.1340, -17.7848, -18.3971]], [[-12.6072, -14.9486, -14.6631], [-13.7629, -17.0907, -17.7745], [-12.7899, -16.1695, -17.1671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": __UpperCAmelCase : Optional[int] = torch.tensor( [ [[-11.9295, -13.4057, -14.8106], [-13.3431, -14.8179, -15.3781], [-14.2836, -15.5942, -16.1588]], [[-11.4906, -12.8067, -13.6564], [-13.1189, -14.0500, -14.1543], [-13.8748, -14.5136, -14.8789]], [[0.5374, 0.1067, -0.4742], [0.1141, -0.2255, -0.7099], [-0.3000, -0.5924, -1.3105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": __UpperCAmelCase : Optional[Any] = torch.tensor( [ [[-7.8217, -9.8767, -10.1717], [-9.4438, -10.9058, -11.4047], [-9.7939, -12.3495, -12.1079]], [[-7.1514, -9.5336, -10.0860], [-9.7776, -11.6822, -11.8439], [-10.1411, -12.7655, -12.8972]], [[0.3021, 0.0805, -0.2310], [-0.0328, -0.1605, -0.2714], [-0.1408, -0.5477, -0.6976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": __UpperCAmelCase : Any = torch.tensor( [ [ [-1.1_3_7_2e0_1, -1.2_7_8_7e0_1, -1.3_4_7_7e0_1], [-1.2_5_3_6e0_1, -1.4_1_9_4e0_1, -1.4_4_0_9e0_1], [-1.3_2_1_7e0_1, -1.4_8_8_8e0_1, -1.5_3_2_7e0_1], ], [ [-1.4_7_9_1e0_1, -1.7_1_2_2e0_1, -1.8_2_7_7e0_1], [-1.7_1_6_3e0_1, -1.9_1_9_2e0_1, -1.9_5_3_3e0_1], [-1.7_8_9_7e0_1, -1.9_9_9_1e0_1, -2.0_3_1_5e0_1], ], [ [7.6_7_2_3e-0_1, 4.1_9_2_1e-0_1, -7.7_8_7_8e-0_2], [4.7_7_7_2e-0_1, 9.5_5_5_7e-0_3, -2.8_0_8_2e-0_1], [3.6_0_3_2e-0_1, -2.4_8_2_6e-0_1, -5.1_1_6_8e-0_1], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": __UpperCAmelCase : Optional[Any] = torch.tensor( [ [[-9.4959, -11.3087, -11.7479], [-11.0025, -12.6540, -12.3319], [-11.4064, -13.0487, -12.9905]], [[-9.8905, -11.3084, -12.0854], [-11.1726, -12.7698, -12.9583], [-11.5985, -13.3278, -14.1774]], [[0.2213, 0.0192, -0.2466], [-0.1731, -0.4213, -0.4874], [-0.3126, -0.6541, -1.1389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": __UpperCAmelCase : Any = torch.tensor( [ [[-13.5748, -13.9111, -12.6500], [-14.3500, -15.3683, -14.2328], [-14.7532, -16.0424, -15.6087]], [[-17.1651, -15.8725, -12.9653], [-17.2580, -17.3718, -14.8223], [-16.6058, -16.8783, -16.7452]], [[-3.6456, -3.0209, -1.4203], [-3.0797, -3.1959, -2.0000], [-1.8757, -1.9217, -1.6997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": __UpperCAmelCase : Dict = torch.tensor( [ [[-16.0976, -16.4856, -17.3962], [-16.6234, -19.0342, -19.7685], [-16.0900, -18.0661, -19.1180]], [[-18.4750, -18.8488, -19.5074], [-19.4030, -22.1570, -22.5977], [-19.1191, -20.8486, -22.3783]], [[-4.5178, -5.5037, -6.5109], [-5.0884, -7.2174, -8.0334], [-4.4156, -5.8117, -7.2970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": __UpperCAmelCase : Optional[Any] = torch.tensor( [ [[-14.2081, -14.4732, -14.1977], [-14.5867, -16.4423, -16.6356], [-13.4441, -14.9685, -16.8696]], [[-14.4576, -14.7073, -15.0451], [-15.0816, -17.6237, -17.9873], [-14.4213, -16.0199, -18.5992]], [[-4.7349, -4.9588, -5.0966], [-4.3210, -6.9325, -7.2591], [-3.4312, -4.7484, -7.1917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": __UpperCAmelCase : str = torch.tensor( [ [[-11.7737, -11.9526, -11.3273], [-13.6692, -14.4574, -13.8878], [-13.8937, -14.6924, -15.9345]], [[-14.6706, -14.5330, -14.1306], [-16.1502, -16.8180, -16.4269], [-16.8338, -17.8939, -20.1746]], [[1.0491, 0.8289, 1.0310], [1.1044, 0.5219, 0.8055], [1.0899, 0.6926, 0.5590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": __UpperCAmelCase : Tuple = torch.tensor( [ [[-12.5641, -13.4777, -13.0684], [-13.9587, -15.8983, -16.6557], [-13.3109, -15.7350, -16.3141]], [[-14.7074, -15.4352, -14.5944], [-16.6353, -18.1663, -18.6120], [-15.1702, -18.0329, -18.1547]], [[-1.7990, -2.0951, -1.7784], [-2.6397, -3.8245, -3.9686], [-1.5264, -2.8126, -2.9316]], ] ) else: __UpperCAmelCase : Dict = logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , UpperCamelCase , atol=1e-2 ) # finally, save model and image processor logger.info(f"Saving PyTorch model and image processor to {pytorch_dump_folder_path}..." ) Path(UpperCamelCase ).mkdir(exist_ok=UpperCamelCase ) model.save_pretrained(UpperCamelCase ) image_processor.save_pretrained(UpperCamelCase ) if __name__ == "__main__": A = argparse.ArgumentParser() parser.add_argument( """--model_name""", default="""segformer.b0.512x512.ade.160k""", type=str, help="""Name of the model you'd like to convert.""", ) parser.add_argument( """--checkpoint_path""", default=None, type=str, help="""Path to the original PyTorch checkpoint (.pth file).""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) A = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
77
"""simple docstring""" import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) class a__ ( __magic_name__ ): lowercase_ = ["input_ids", "attention_mask"] def __init__( self : Optional[Any] , UpperCamelCase_ : List[Any]="</s>" , UpperCamelCase_ : Tuple="<unk>" , UpperCamelCase_ : List[str]="<pad>" , UpperCamelCase_ : Union[str, Any]=125 , UpperCamelCase_ : Dict=None , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: __UpperCAmelCase : int = [F"<extra_id_{i}>" for i in range(UpperCamelCase_)] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens __UpperCAmelCase : Dict = len(set(filter(lambda UpperCamelCase_: bool("extra_id" in str(UpperCamelCase_)) , UpperCamelCase_))) if extra_tokens != extra_ids: raise ValueError( F"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" " provided to ByT5Tokenizer. In this case the additional_special_tokens must include the" " extra_ids tokens") __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else pad_token __UpperCAmelCase : List[str] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else eos_token __UpperCAmelCase : Optional[int] = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_) if isinstance(UpperCamelCase_ , UpperCamelCase_) else unk_token super().__init__( eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , extra_ids=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : List[str] = extra_ids __UpperCAmelCase : int = 2**8 # utf is 8 bits # define special tokens dict __UpperCAmelCase : Dict[int, str] = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } __UpperCAmelCase : Any = len(self.special_tokens_encoder) __UpperCAmelCase : List[Any] = len(UpperCamelCase_) for i, token in enumerate(UpperCamelCase_): __UpperCAmelCase : Union[str, Any] = self.vocab_size + i - n __UpperCAmelCase : Dict[str, int] = {v: k for k, v in self.special_tokens_encoder.items()} @property def a_ ( self : List[Any]): """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def a_ ( self : List[str] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_) # normal case: some special tokens if token_ids_a is None: return ([0] * len(UpperCamelCase_)) + [1] return ([0] * len(UpperCamelCase_)) + [1] + ([0] * len(UpperCamelCase_)) + [1] def a_ ( self : Optional[Any] , UpperCamelCase_ : List[int]): """simple docstring""" if len(UpperCamelCase_) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F"This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated" " eos tokens being added.") return token_ids else: return token_ids + [self.eos_token_id] def a_ ( self : Dict , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Dict = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos) * [0] return len(token_ids_a + eos + token_ids_a + eos) * [0] def a_ ( self : Optional[int] , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" __UpperCAmelCase : Optional[Any] = self._add_eos_if_not_present(UpperCamelCase_) if token_ids_a is None: return token_ids_a else: __UpperCAmelCase : List[Any] = self._add_eos_if_not_present(UpperCamelCase_) return token_ids_a + token_ids_a def a_ ( self : List[str] , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : Any = [chr(UpperCamelCase_) for i in text.encode("utf-8")] return tokens def a_ ( self : Tuple , UpperCamelCase_ : List[Any]): """simple docstring""" if token in self.special_tokens_encoder: __UpperCAmelCase : Any = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: __UpperCAmelCase : int = self.added_tokens_encoder[token] elif len(UpperCamelCase_) != 1: __UpperCAmelCase : Optional[Any] = self.unk_token_id else: __UpperCAmelCase : Any = ord(UpperCamelCase_) + self._num_special_tokens return token_id def a_ ( self : Any , UpperCamelCase_ : List[str]): """simple docstring""" if index in self.special_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[index] else: __UpperCAmelCase : List[str] = chr(index - self._num_special_tokens) return token def a_ ( self : Dict , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : str = b"" for token in tokens: if token in self.special_tokens_decoder: __UpperCAmelCase : Tuple = self.special_tokens_decoder[token].encode("utf-8") elif token in self.added_tokens_decoder: __UpperCAmelCase : Any = self.special_tokens_decoder[token].encode("utf-8") elif token in self.special_tokens_encoder: __UpperCAmelCase : Optional[int] = token.encode("utf-8") elif token in self.added_tokens_encoder: __UpperCAmelCase : Optional[Any] = token.encode("utf-8") else: __UpperCAmelCase : Any = bytes([ord(UpperCamelCase_)]) bstring += tok_string __UpperCAmelCase : List[Any] = bstring.decode("utf-8" , errors="ignore") return string def a_ ( self : Optional[int] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" return ()
77
1
"""simple docstring""" import logging import os from typing import List, Tuple import numpy as np import psutil import torch import torch.distributed as dist from transformers import RagRetriever A = logging.getLogger(__name__) class a__ ( __magic_name__ ): def __init__( self : Tuple , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Any=None): """simple docstring""" super().__init__( UpperCamelCase_ , question_encoder_tokenizer=UpperCamelCase_ , generator_tokenizer=UpperCamelCase_ , index=UpperCamelCase_ , init_retrieval=UpperCamelCase_ , ) __UpperCAmelCase : Dict = None def a_ ( self : Optional[int] , UpperCamelCase_ : int): """simple docstring""" logger.info("initializing retrieval") # initializing a separate process group for retrieval as the default # nccl backend doesn't support gather/scatter operations while gloo # is too slow to replace nccl for the core gpu communication if dist.is_initialized(): logger.info("dist initialized") # needs to be set manually __UpperCAmelCase : Optional[Any] = self._infer_socket_ifname() # avoid clash with the NCCL port __UpperCAmelCase : Any = str(distributed_port + 1) __UpperCAmelCase : Optional[Any] = dist.new_group(ranks=UpperCamelCase_ , backend="gloo") # initialize retriever only on the main worker if not dist.is_initialized() or self._is_main(): logger.info("dist not initialized / main") self.index.init_index() # all processes wait untill the retriever is initialized by the main process if dist.is_initialized(): torch.distributed.barrier(group=self.process_group) def a_ ( self : List[str]): """simple docstring""" return dist.get_rank(group=self.process_group) == 0 def a_ ( self : List[str] , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Dict=torch.floataa): """simple docstring""" __UpperCAmelCase : Optional[Any] = torch.empty(UpperCamelCase_ , dtype=UpperCamelCase_) dist.scatter(UpperCamelCase_ , src=0 , scatter_list=UpperCamelCase_ , group=self.process_group) return target_tensor def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = psutil.net_if_addrs() # a hacky way to deal with varying network interface names __UpperCAmelCase : Dict = next((addr for addr in addrs if addr.startswith("e")) , UpperCamelCase_) return ifname def a_ ( self : Union[str, Any] , UpperCamelCase_ : np.ndarray , UpperCamelCase_ : int): """simple docstring""" if not dist.is_initialized(): __UpperCAmelCase , __UpperCAmelCase : str = self._main_retrieve(UpperCamelCase_ , UpperCamelCase_) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(UpperCamelCase_) # distributed training __UpperCAmelCase : Union[str, Any] = dist.get_world_size(group=self.process_group) # gather logic __UpperCAmelCase : List[Any] = None if self._is_main(): __UpperCAmelCase : str = [torch.empty(question_hidden_states.shape , dtype=torch.floataa) for _ in range(UpperCamelCase_)] dist.gather(torch.tensor(UpperCamelCase_) , dst=0 , gather_list=UpperCamelCase_ , group=self.process_group) # scatter logic __UpperCAmelCase : Tuple = question_hidden_states.shape[0] __UpperCAmelCase : List[Any] = [] __UpperCAmelCase : Union[str, Any] = [] if self._is_main(): assert len(UpperCamelCase_) == world_size __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = self._main_retrieve(torch.cat(UpperCamelCase_).numpy() , UpperCamelCase_) __UpperCAmelCase , __UpperCAmelCase : int = torch.tensor(UpperCamelCase_), torch.tensor(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = self._chunk_tensor(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : int = self._chunk_tensor(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[Any] = self._scattered(UpperCamelCase_ , [n_queries, n_docs] , target_type=torch.intaa) __UpperCAmelCase : Optional[int] = self._scattered(UpperCamelCase_ , [n_queries, n_docs, question_hidden_states.shape[1]]) return retrieved_doc_embeds.numpy(), doc_ids.numpy(), self.index.get_doc_dicts(UpperCamelCase_)
77
"""simple docstring""" import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a__ ( unittest.TestCase ): def __init__( self : List[Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Tuple=3 , UpperCamelCase_ : Optional[int]=32 , UpperCamelCase_ : Dict=3 , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : str=[10, 20, 30, 40] , UpperCamelCase_ : Tuple=[1, 1, 2, 1] , UpperCamelCase_ : str=True , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Dict="relu" , UpperCamelCase_ : str=3 , UpperCamelCase_ : int=None , ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : List[str] = image_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : Union[str, Any] = embeddings_size __UpperCAmelCase : Dict = hidden_sizes __UpperCAmelCase : Dict = depths __UpperCAmelCase : Tuple = is_training __UpperCAmelCase : List[Any] = use_labels __UpperCAmelCase : Optional[int] = hidden_act __UpperCAmelCase : str = num_labels __UpperCAmelCase : Optional[int] = scope __UpperCAmelCase : Dict = len(UpperCamelCase_) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) __UpperCAmelCase : Dict = self.get_config() return config, pixel_values def a_ ( self : Dict): """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def a_ ( self : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[str] = FlaxRegNetModel(config=UpperCamelCase_) __UpperCAmelCase : Dict = model(UpperCamelCase_) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def a_ ( self : Any , UpperCamelCase_ : Dict , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : List[Any] = self.num_labels __UpperCAmelCase : Tuple = FlaxRegNetForImageClassification(config=UpperCamelCase_) __UpperCAmelCase : str = model(UpperCamelCase_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Any = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase : Tuple = config_and_inputs __UpperCAmelCase : Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () lowercase_ = False lowercase_ = False lowercase_ = False def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = FlaxRegNetModelTester(self) __UpperCAmelCase : str = ConfigTester(self , config_class=UpperCamelCase_ , has_text_modality=UpperCamelCase_) def a_ ( self : Dict): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def a_ ( self : Tuple): """simple docstring""" return def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase_) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase_) @unittest.skip(reason="RegNet does not use inputs_embeds") def a_ ( self : Union[str, Any]): """simple docstring""" pass @unittest.skip(reason="RegNet does not support input and output embeddings") def a_ ( self : Optional[int]): """simple docstring""" pass def a_ ( self : str): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : int = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[int] = inspect.signature(model.__call__) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : Any = [*signature.parameters.keys()] __UpperCAmelCase : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase_) def a_ ( self : int): """simple docstring""" def check_hidden_states_output(UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Tuple , UpperCamelCase_ : Union[str, Any]): __UpperCAmelCase : Union[str, Any] = model_class(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = model(**self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_)) __UpperCAmelCase : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCAmelCase : str = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase_) , expected_num_stages + 1) __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : List[str] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase : Optional[int] = True check_hidden_states_output(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__): __UpperCAmelCase : List[Any] = self._prepare_for_class(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[int] = model_class(UpperCamelCase_) @jax.jit def model_jitted(UpperCamelCase_ : int , **UpperCamelCase_ : Optional[int]): return model(pixel_values=UpperCamelCase_ , **UpperCamelCase_) with self.subTest("JIT Enabled"): __UpperCAmelCase : Optional[Any] = model_jitted(**UpperCamelCase_).to_tuple() with self.subTest("JIT Disabled"): with jax.disable_jit(): __UpperCAmelCase : Dict = model_jitted(**UpperCamelCase_).to_tuple() self.assertEqual(len(UpperCamelCase_) , len(UpperCamelCase_)) for jitted_output, output in zip(UpperCamelCase_ , UpperCamelCase_): self.assertEqual(jitted_output.shape , output.shape) def _UpperCamelCase ( ) -> Any: """simple docstring""" __UpperCAmelCase : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_flax class a__ ( unittest.TestCase ): @cached_property def a_ ( self : Optional[int]): """simple docstring""" return AutoImageProcessor.from_pretrained("facebook/regnet-y-040") if is_vision_available() else None @slow def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Any = FlaxRegNetForImageClassification.from_pretrained("facebook/regnet-y-040") __UpperCAmelCase : Dict = self.default_image_processor __UpperCAmelCase : str = prepare_img() __UpperCAmelCase : int = image_processor(images=UpperCamelCase_ , return_tensors="np") __UpperCAmelCase : Dict = model(**UpperCamelCase_) # verify the logits __UpperCAmelCase : Dict = (1, 1000) self.assertEqual(outputs.logits.shape , UpperCamelCase_) __UpperCAmelCase : Any = jnp.array([-0.4180, -1.5051, -3.4836]) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , UpperCamelCase_ , atol=1e-4))
77
1
"""simple docstring""" import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Dict , UpperCamelCase_ : str , UpperCamelCase_ : List[Any]=7 , UpperCamelCase_ : Any=3 , UpperCamelCase_ : Dict=18 , UpperCamelCase_ : List[str]=30 , UpperCamelCase_ : Tuple=400 , UpperCamelCase_ : Dict=True , UpperCamelCase_ : Union[str, Any]=None , UpperCamelCase_ : Tuple=True , ): """simple docstring""" __UpperCAmelCase : Any = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : List[str] = parent __UpperCAmelCase : Optional[Any] = batch_size __UpperCAmelCase : List[Any] = num_channels __UpperCAmelCase : str = image_size __UpperCAmelCase : Tuple = min_resolution __UpperCAmelCase : Optional[Any] = max_resolution __UpperCAmelCase : Optional[Any] = do_resize __UpperCAmelCase : int = size __UpperCAmelCase : List[str] = do_normalize def a_ ( self : List[Any]): """simple docstring""" return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8866443634033203, 0.6618829369544983, 0.3891746401786804], [-0.6042559146881104, -0.02295008860528469, 0.5423797369003296], ]), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ImageGPTImageProcessor if is_vision_available() else None def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[int] = ImageGPTImageProcessingTester(self) @property def a_ ( self : Any): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "clusters")) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Dict = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {"height": 18, "width": 18}) __UpperCAmelCase : Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict , size=42) self.assertEqual(image_processor.size , {"height": 42, "width": 42}) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict) __UpperCAmelCase : Optional[int] = json.loads(image_processor.to_json_string()) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(UpperCamelCase_ , obj[key])) else: self.assertEqual(obj[key] , UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) with tempfile.TemporaryDirectory() as tmpdirname: __UpperCAmelCase : Any = os.path.join(UpperCamelCase_ , "image_processor.json") image_processor_first.to_json_file(UpperCamelCase_) __UpperCAmelCase : Tuple = self.image_processing_class.from_json_file(UpperCamelCase_).to_dict() __UpperCAmelCase : List[Any] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(UpperCamelCase_ , image_processor_second[key])) else: self.assertEqual(image_processor_first[key] , UpperCamelCase_) def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = self.image_processing_class.from_pretrained(UpperCamelCase_).to_dict() __UpperCAmelCase : List[Any] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(UpperCamelCase_ , image_processor_second[key])) else: self.assertEqual(image_processor_first[key] , UpperCamelCase_) @unittest.skip("ImageGPT requires clusters at initialization") def a_ ( self : Tuple): """simple docstring""" pass def _UpperCamelCase ( ) -> Tuple: """simple docstring""" __UpperCAmelCase : Union[str, Any] = load_dataset("hf-internal-testing/fixtures_image_utils" , split="test" ) __UpperCAmelCase : List[str] = Image.open(dataset[4]["file"] ) __UpperCAmelCase : Union[str, Any] = Image.open(dataset[5]["file"] ) __UpperCAmelCase : Tuple = [imagea, imagea] return images @require_vision @require_torch class a__ ( unittest.TestCase ): @slow def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Tuple = ImageGPTImageProcessor.from_pretrained("openai/imagegpt-small") __UpperCAmelCase : Any = prepare_images() # test non-batched __UpperCAmelCase : List[Any] = image_processing(images[0] , return_tensors="pt") self.assertIsInstance(encoding.input_ids , torch.LongTensor) self.assertEqual(encoding.input_ids.shape , (1, 1024)) __UpperCAmelCase : Tuple = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , UpperCamelCase_) # test batched __UpperCAmelCase : List[Any] = image_processing(UpperCamelCase_ , return_tensors="pt") self.assertIsInstance(encoding.input_ids , torch.LongTensor) self.assertEqual(encoding.input_ids.shape , (2, 1024)) __UpperCAmelCase : int = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , UpperCamelCase_)
77
"""simple docstring""" from scipy.stats import spearmanr import datasets A = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ A = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ A = r"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): def a_ ( self : Any): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float"), "references": datasets.Value("float"), }) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def a_ ( self : Union[str, Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : int=False): """simple docstring""" __UpperCAmelCase : List[str] = spearmanr(UpperCamelCase_ , UpperCamelCase_) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
77
1
"""simple docstring""" import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) A = logging.getLogger() def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : Optional[int] = {} __UpperCAmelCase : str = os.path.join(UpperCamelCase , "all_results.json" ) if os.path.exists(UpperCamelCase ): with open(UpperCamelCase , "r" ) as f: __UpperCAmelCase : List[str] = json.load(UpperCamelCase ) else: raise ValueError(f"can't find {path}" ) return results A = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class a__ ( __magic_name__ ): def a_ ( self : Optional[int]): """simple docstring""" import xla_spawn __UpperCAmelCase : Tuple = self.get_auto_remove_tmp_dir() __UpperCAmelCase : List[Any] = F"\n ./examples/pytorch/text-classification/run_glue.py\n --num_cores=8\n ./examples/pytorch/text-classification/run_glue.py\n --model_name_or_path distilbert-base-uncased\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --train_file ./tests/fixtures/tests_samples/MRPC/train.csv\n --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv\n --do_train\n --do_eval\n --debug tpu_metrics_debug\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --learning_rate=1e-4\n --max_steps=10\n --warmup_steps=2\n --seed=42\n --max_seq_length=128\n ".split() with patch.object(UpperCamelCase_ , "argv" , UpperCamelCase_): __UpperCAmelCase : str = time() xla_spawn.main() __UpperCAmelCase : Tuple = time() __UpperCAmelCase : Optional[Any] = get_results(UpperCamelCase_) self.assertGreaterEqual(result["eval_accuracy"] , 0.75) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500) def a_ ( self : Optional[int]): """simple docstring""" import xla_spawn __UpperCAmelCase : Tuple = "\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n ".split() with patch.object(UpperCamelCase_ , "argv" , UpperCamelCase_): xla_spawn.main()
77
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) A = {"""vocab_file""": """spiece.model"""} A = { """vocab_file""": { """bert_for_seq_generation""": ( """https://huggingface.co/google/bert_for_seq_generation_L-24_bbc_encoder/resolve/main/spiece.model""" ), } } A = {"""bert_for_seq_generation""": 512} class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = [] lowercase_ = ["input_ids", "attention_mask"] def __init__( self : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : List[str]="<s>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Optional[int]="<unk>" , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : List[Any]="<::::>" , UpperCamelCase_ : Optional[Dict[str, Any]] = None , **UpperCamelCase_ : List[Any] , ): """simple docstring""" __UpperCAmelCase : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs # Add extra_ids to the special token list super().__init__( bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) __UpperCAmelCase : Dict = vocab_file __UpperCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(UpperCamelCase_) @property def a_ ( self : List[str]): """simple docstring""" return self.sp_model.get_piece_size() def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : int = {self.convert_ids_to_tokens(UpperCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self : int): """simple docstring""" __UpperCAmelCase : Optional[int] = self.__dict__.copy() __UpperCAmelCase : List[Any] = None return state def __setstate__( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): __UpperCAmelCase : List[Any] = {} __UpperCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def a_ ( self : Any , UpperCamelCase_ : str): """simple docstring""" return self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_) def a_ ( self : Optional[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" return self.sp_model.piece_to_id(UpperCamelCase_) def a_ ( self : Tuple , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : int = self.sp_model.IdToPiece(UpperCamelCase_) return token def a_ ( self : Dict , UpperCamelCase_ : Optional[Any]): """simple docstring""" __UpperCAmelCase : int = [] __UpperCAmelCase : Tuple = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCamelCase_) + token __UpperCAmelCase : List[Any] = [] else: current_sub_tokens.append(UpperCamelCase_) out_string += self.sp_model.decode(UpperCamelCase_) return out_string.strip() def a_ ( self : List[str] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : Tuple = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , UpperCamelCase_) elif not os.path.isfile(self.vocab_file): with open(UpperCamelCase_ , "wb") as fi: __UpperCAmelCase : List[str] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A = logging.get_logger(__name__) A = { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/config.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/config.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json""" ), """distilbert-base-uncased-finetuned-sst-2-english""": ( """https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json""" ), } class a__ ( __magic_name__ ): lowercase_ = "distilbert" lowercase_ = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self : Optional[int] , UpperCamelCase_ : str=30522 , UpperCamelCase_ : Tuple=512 , UpperCamelCase_ : Optional[int]=False , UpperCamelCase_ : List[str]=6 , UpperCamelCase_ : str=12 , UpperCamelCase_ : List[Any]=768 , UpperCamelCase_ : Optional[Any]=4 * 768 , UpperCamelCase_ : int=0.1 , UpperCamelCase_ : Any=0.1 , UpperCamelCase_ : Tuple="gelu" , UpperCamelCase_ : Any=0.02 , UpperCamelCase_ : Optional[int]=0.1 , UpperCamelCase_ : Tuple=0.2 , UpperCamelCase_ : Optional[int]=0 , **UpperCamelCase_ : List[str] , ): """simple docstring""" __UpperCAmelCase : int = vocab_size __UpperCAmelCase : Tuple = max_position_embeddings __UpperCAmelCase : str = sinusoidal_pos_embds __UpperCAmelCase : Union[str, Any] = n_layers __UpperCAmelCase : str = n_heads __UpperCAmelCase : Union[str, Any] = dim __UpperCAmelCase : str = hidden_dim __UpperCAmelCase : Optional[int] = dropout __UpperCAmelCase : Optional[int] = attention_dropout __UpperCAmelCase : Optional[int] = activation __UpperCAmelCase : int = initializer_range __UpperCAmelCase : int = qa_dropout __UpperCAmelCase : Optional[Any] = seq_classif_dropout super().__init__(**UpperCamelCase_ , pad_token_id=UpperCamelCase_) class a__ ( __magic_name__ ): @property def a_ ( self : Union[str, Any]): """simple docstring""" if self.task == "multiple-choice": __UpperCAmelCase : Dict = {0: "batch", 1: "choice", 2: "sequence"} else: __UpperCAmelCase : Optional[int] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ])
77
"""simple docstring""" import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed A = """true""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=16 ) -> Tuple: """simple docstring""" set_seed(42 ) __UpperCAmelCase : Dict = RegressionModel() __UpperCAmelCase : Optional[Any] = deepcopy(UpperCamelCase ) __UpperCAmelCase : Any = RegressionDataset(length=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = DataLoader(UpperCamelCase , batch_size=UpperCamelCase ) model.to(accelerator.device ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return model, ddp_model, dataloader def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=False ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[str] = AutoTokenizer.from_pretrained("hf-internal-testing/mrpc-bert-base-cased" ) __UpperCAmelCase : Dict = load_dataset("glue" , "mrpc" , split="validation" ) def tokenize_function(UpperCamelCase ): __UpperCAmelCase : Dict = tokenizer(examples["sentence1"] , examples["sentence2"] , truncation=UpperCamelCase , max_length=UpperCamelCase ) return outputs with accelerator.main_process_first(): __UpperCAmelCase : str = dataset.map( UpperCamelCase , batched=UpperCamelCase , remove_columns=["idx", "sentence1", "sentence2"] , ) __UpperCAmelCase : List[Any] = tokenized_datasets.rename_column("label" , "labels" ) def collate_fn(UpperCamelCase ): if use_longest: return tokenizer.pad(UpperCamelCase , padding="longest" , return_tensors="pt" ) return tokenizer.pad(UpperCamelCase , padding="max_length" , max_length=128 , return_tensors="pt" ) return DataLoader(UpperCamelCase , shuffle=UpperCamelCase , collate_fn=UpperCamelCase , batch_size=16 ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Optional[int]: """simple docstring""" __UpperCAmelCase : List[Any] = Accelerator(dispatch_batches=UpperCamelCase , split_batches=UpperCamelCase ) __UpperCAmelCase : int = get_dataloader(UpperCamelCase , not dispatch_batches ) __UpperCAmelCase : Any = AutoModelForSequenceClassification.from_pretrained( "hf-internal-testing/mrpc-bert-base-cased" , return_dict=UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Dict = accelerator.prepare(UpperCamelCase , UpperCamelCase ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Dict = [] for batch in dataloader: __UpperCAmelCase , __UpperCAmelCase : int = batch.values() with torch.no_grad(): __UpperCAmelCase : int = model(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : List[str] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = [], [] for logit, targ in logits_and_targets: logits.append(UpperCamelCase ) targs.append(UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = torch.cat(UpperCamelCase ), torch.cat(UpperCamelCase ) return logits, targs def _UpperCamelCase ( UpperCamelCase , UpperCamelCase=82 , UpperCamelCase=False , UpperCamelCase=False , UpperCamelCase=16 ) -> int: """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = get_basic_setup(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Union[str, Any] = generate_predictions(UpperCamelCase , UpperCamelCase , UpperCamelCase ) assert ( len(UpperCamelCase ) == num_samples ), f"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(UpperCamelCase )}" def _UpperCamelCase ( UpperCamelCase = False , UpperCamelCase = False ) -> List[str]: """simple docstring""" __UpperCAmelCase : List[str] = evaluate.load("glue" , "mrpc" ) __UpperCAmelCase , __UpperCAmelCase : List[Any] = get_mrpc_setup(UpperCamelCase , UpperCamelCase ) # First do baseline __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = setup["no"] model.to(UpperCamelCase ) model.eval() for batch in dataloader: batch.to(UpperCamelCase ) with torch.inference_mode(): __UpperCAmelCase : List[str] = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=UpperCamelCase , references=batch["labels"] ) __UpperCAmelCase : str = metric.compute() # Then do distributed __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = setup["ddp"] model.eval() for batch in dataloader: with torch.inference_mode(): __UpperCAmelCase : Any = model(**UpperCamelCase ) __UpperCAmelCase : str = outputs.logits.argmax(dim=-1 ) __UpperCAmelCase : Union[str, Any] = batch["labels"] __UpperCAmelCase , __UpperCAmelCase : Any = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=UpperCamelCase , references=UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def _UpperCamelCase ( ) -> List[Any]: """simple docstring""" __UpperCAmelCase : Dict = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print("**Testing gather_for_metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(UpperCamelCase , UpperCamelCase ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test torch metrics**" ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __UpperCAmelCase : Union[str, Any] = Accelerator(split_batches=UpperCamelCase , dispatch_batches=UpperCamelCase ) if accelerator.is_local_main_process: print(f"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(UpperCamelCase , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print("**Test last batch is not dropped when perfectly divisible**" ) __UpperCAmelCase : Any = Accelerator() test_torch_metrics(UpperCamelCase , 512 ) accelerator.state._reset_state() def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
77
1
"""simple docstring""" import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) A = { """iou_prediction_head.layers.0""": """iou_prediction_head.proj_in""", """iou_prediction_head.layers.1""": """iou_prediction_head.layers.0""", """iou_prediction_head.layers.2""": """iou_prediction_head.proj_out""", """mask_decoder.output_upscaling.0""": """mask_decoder.upscale_conv1""", """mask_decoder.output_upscaling.1""": """mask_decoder.upscale_layer_norm""", """mask_decoder.output_upscaling.3""": """mask_decoder.upscale_conv2""", """mask_downscaling.0""": """mask_embed.conv1""", """mask_downscaling.1""": """mask_embed.layer_norm1""", """mask_downscaling.3""": """mask_embed.conv2""", """mask_downscaling.4""": """mask_embed.layer_norm2""", """mask_downscaling.6""": """mask_embed.conv3""", """point_embeddings""": """point_embed""", """pe_layer.positional_encoding_gaussian_matrix""": """shared_embedding.positional_embedding""", """image_encoder""": """vision_encoder""", """neck.0""": """neck.conv1""", """neck.1""": """neck.layer_norm1""", """neck.2""": """neck.conv2""", """neck.3""": """neck.layer_norm2""", """patch_embed.proj""": """patch_embed.projection""", """.norm""": """.layer_norm""", """blocks""": """layers""", } def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : Optional[Any] = {} state_dict.pop("pixel_mean" , UpperCamelCase ) state_dict.pop("pixel_std" , UpperCamelCase ) __UpperCAmelCase : Any = R".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: __UpperCAmelCase : Union[str, Any] = key.replace(UpperCamelCase , UpperCamelCase ) if re.match(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : int = int(re.match(UpperCamelCase , UpperCamelCase ).group(2 ) ) if layer_nb == 0: __UpperCAmelCase : Dict = key.replace("layers.0" , "proj_in" ) elif layer_nb == 1: __UpperCAmelCase : int = key.replace("layers.1" , "layers.0" ) elif layer_nb == 2: __UpperCAmelCase : Union[str, Any] = key.replace("layers.2" , "proj_out" ) __UpperCAmelCase : Tuple = value __UpperCAmelCase : Any = model_state_dict[ "prompt_encoder.shared_embedding.positional_embedding" ] return model_state_dict def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase="ybelkada/segment-anything" ) -> Dict: """simple docstring""" __UpperCAmelCase : int = hf_hub_download(UpperCamelCase , f"checkpoints/{model_name}.pth" ) if "sam_vit_b" in model_name: __UpperCAmelCase : List[Any] = SamConfig() elif "sam_vit_l" in model_name: __UpperCAmelCase : int = SamVisionConfig( hidden_size=1024 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) __UpperCAmelCase : Dict = SamConfig( vision_config=UpperCamelCase , ) elif "sam_vit_h" in model_name: __UpperCAmelCase : str = SamVisionConfig( hidden_size=1280 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) __UpperCAmelCase : List[Any] = SamConfig( vision_config=UpperCamelCase , ) __UpperCAmelCase : str = torch.load(UpperCamelCase , map_location="cpu" ) __UpperCAmelCase : Optional[Any] = replace_keys(UpperCamelCase ) __UpperCAmelCase : Tuple = SamImageProcessor() __UpperCAmelCase : Optional[Any] = SamProcessor(image_processor=UpperCamelCase ) __UpperCAmelCase : List[Any] = SamModel(UpperCamelCase ) hf_model.load_state_dict(UpperCamelCase ) __UpperCAmelCase : Any = hf_model.to("cuda" ) __UpperCAmelCase : Tuple = "https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png" __UpperCAmelCase : int = Image.open(requests.get(UpperCamelCase , stream=UpperCamelCase ).raw ).convert("RGB" ) __UpperCAmelCase : List[Any] = [[[400, 650]]] __UpperCAmelCase : str = [[1]] __UpperCAmelCase : List[Any] = processor(images=np.array(UpperCamelCase ) , return_tensors="pt" ).to("cuda" ) with torch.no_grad(): __UpperCAmelCase : int = hf_model(**UpperCamelCase ) __UpperCAmelCase : int = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579890251159668 __UpperCAmelCase : str = processor( images=np.array(UpperCamelCase ) , input_points=UpperCamelCase , input_labels=UpperCamelCase , return_tensors="pt" ).to("cuda" ) with torch.no_grad(): __UpperCAmelCase : List[Any] = hf_model(**UpperCamelCase ) __UpperCAmelCase : int = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712603092193604 __UpperCAmelCase : Dict = ((75, 275, 1725, 850),) __UpperCAmelCase : List[Any] = processor(images=np.array(UpperCamelCase ) , input_boxes=UpperCamelCase , return_tensors="pt" ).to("cuda" ) with torch.no_grad(): __UpperCAmelCase : Tuple = hf_model(**UpperCamelCase ) __UpperCAmelCase : int = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686015605926514 # Test with 2 points and 1 image. __UpperCAmelCase : List[str] = [[[400, 650], [800, 650]]] __UpperCAmelCase : int = [[1, 1]] __UpperCAmelCase : List[Any] = processor( images=np.array(UpperCamelCase ) , input_points=UpperCamelCase , input_labels=UpperCamelCase , return_tensors="pt" ).to("cuda" ) with torch.no_grad(): __UpperCAmelCase : Optional[int] = hf_model(**UpperCamelCase ) __UpperCAmelCase : Optional[Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936047792434692 if __name__ == "__main__": A = argparse.ArgumentParser() A = ["""sam_vit_b_01ec64""", """sam_vit_h_4b8939""", """sam_vit_l_0b3195"""] parser.add_argument( """--model_name""", default="""sam_vit_h_4b8939""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub after converting""", ) parser.add_argument( """--model_hub_id""", default="""ybelkada/segment-anything""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) A = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
77
"""simple docstring""" import math def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = 0 , UpperCamelCase = 0 ) -> list: """simple docstring""" __UpperCAmelCase : Union[str, Any] = end or len(UpperCamelCase ) for i in range(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[Any] = i __UpperCAmelCase : Any = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: __UpperCAmelCase : Dict = array[temp_index - 1] temp_index -= 1 __UpperCAmelCase : str = temp_index_value return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: # Max Heap """simple docstring""" __UpperCAmelCase : Optional[Any] = index __UpperCAmelCase : List[str] = 2 * index + 1 # Left Node __UpperCAmelCase : Union[str, Any] = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: __UpperCAmelCase : Tuple = left_index if right_index < heap_size and array[largest] < array[right_index]: __UpperCAmelCase : int = right_index if largest != index: __UpperCAmelCase , __UpperCAmelCase : List[str] = array[largest], array[index] heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" __UpperCAmelCase : List[Any] = len(UpperCamelCase ) for i in range(n // 2 , -1 , -1 ): heapify(UpperCamelCase , UpperCamelCase , UpperCamelCase ) for i in range(n - 1 , 0 , -1 ): __UpperCAmelCase , __UpperCAmelCase : int = array[0], array[i] heapify(UpperCamelCase , 0 , UpperCamelCase ) return array def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Optional[Any] = low __UpperCAmelCase : List[str] = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i __UpperCAmelCase , __UpperCAmelCase : Optional[int] = array[j], array[i] i += 1 def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" if len(UpperCamelCase ) == 0: return array __UpperCAmelCase : Optional[int] = 2 * math.ceil(math.loga(len(UpperCamelCase ) ) ) __UpperCAmelCase : List[Any] = 16 return intro_sort(UpperCamelCase , 0 , len(UpperCamelCase ) , UpperCamelCase , UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> list: """simple docstring""" while end - start > size_threshold: if max_depth == 0: return heap_sort(UpperCamelCase ) max_depth -= 1 __UpperCAmelCase : List[Any] = median_of_a(UpperCamelCase , UpperCamelCase , start + ((end - start) // 2) + 1 , end - 1 ) __UpperCAmelCase : Union[str, Any] = partition(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) intro_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Optional[Any] = p return insertion_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() A = input("""Enter numbers separated by a comma : """).strip() A = [float(item) for item in user_input.split(""",""")] print(sort(unsorted))
77
1
"""simple docstring""" A = """Input must be a string of 8 numbers plus letter""" A = """TRWAGMYFPDXBNJZSQVHLCKE""" def _UpperCamelCase ( UpperCamelCase ) -> bool: """simple docstring""" if not isinstance(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : List[Any] = f"Expected string as input, found {type(UpperCamelCase ).__name__}" raise TypeError(UpperCamelCase ) __UpperCAmelCase : Union[str, Any] = spanish_id.replace("-" , "" ).upper() if len(UpperCamelCase ) != 9: raise ValueError(UpperCamelCase ) try: __UpperCAmelCase : int = int(spanish_id_clean[0:8] ) __UpperCAmelCase : Optional[Any] = spanish_id_clean[8] except ValueError as ex: raise ValueError(UpperCamelCase ) from ex if letter.isdigit(): raise ValueError(UpperCamelCase ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" import numpy as np from PIL import Image def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : str = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Any = 0 __UpperCAmelCase : Dict = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Tuple = 0 # compute the shape of the output matrix __UpperCAmelCase : Optional[int] = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape maxpool_shape __UpperCAmelCase : List[str] = np.zeros((maxpool_shape, maxpool_shape) ) while i < arr.shape[0]: if i + size > arr.shape[0]: # if the end of the matrix is reached, break break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the maximum of the pooling matrix __UpperCAmelCase : str = np.max(arr[i : i + size, j : j + size] ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : int = 0 __UpperCAmelCase : int = 0 return updated_arr def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> np.ndarray: """simple docstring""" __UpperCAmelCase : List[str] = np.array(UpperCamelCase ) if arr.shape[0] != arr.shape[1]: raise ValueError("The input array is not a square matrix" ) __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : List[str] = 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Any = 0 # compute the shape of the output matrix __UpperCAmelCase : Tuple = (arr.shape[0] - size) // stride + 1 # initialize the output matrix with zeros of shape avgpool_shape __UpperCAmelCase : str = np.zeros((avgpool_shape, avgpool_shape) ) while i < arr.shape[0]: # if the end of the matrix is reached, break if i + size > arr.shape[0]: break while j < arr.shape[1]: # if the end of the matrix is reached, break if j + size > arr.shape[1]: break # compute the average of the pooling matrix __UpperCAmelCase : Tuple = int(np.average(arr[i : i + size, j : j + size] ) ) # shift the pooling matrix by stride of column pixels j += stride mat_j += 1 # shift the pooling matrix by stride of row pixels i += stride mat_i += 1 # reset the column index to 0 __UpperCAmelCase : Tuple = 0 __UpperCAmelCase : Optional[Any] = 0 return updated_arr # Main Function if __name__ == "__main__": from doctest import testmod testmod(name="""avgpooling""", verbose=True) # Loading the image A = Image.open("""path_to_image""") # Converting the image to numpy array and maxpooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(maxpooling(np.array(image), size=3, stride=2)).show() # Converting the image to numpy array and averagepooling, displaying the result # Ensure that the image is a square matrix Image.fromarray(avgpooling(np.array(image), size=3, stride=2)).show()
77
1
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = "cpu" , UpperCamelCase = None ) -> None: """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.load(UpperCamelCase , map_location=UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(UpperCamelCase , torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) __UpperCAmelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __UpperCAmelCase : str = src_path torch.save(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
77
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: A = None A = logging.get_logger(__name__) A = """▁""" A = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} A = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } A = { """google/pegasus-xsum""": 512, } class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = PegasusTokenizer lowercase_ = ["input_ids", "attention_mask"] def __init__( self : str , UpperCamelCase_ : str=None , UpperCamelCase_ : Optional[int]=None , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Any="<unk>" , UpperCamelCase_ : Tuple="<mask_2>" , UpperCamelCase_ : Any="<mask_1>" , UpperCamelCase_ : Tuple=None , UpperCamelCase_ : str=103 , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" __UpperCAmelCase : Optional[int] = offset if additional_special_tokens is not None: if not isinstance(UpperCamelCase_ , UpperCamelCase_): raise TypeError( F"additional_special_tokens should be of type {type(UpperCamelCase_)}, but is" F" {type(UpperCamelCase_)}") __UpperCAmelCase : Any = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"<unk_{i}>" for i in range(len(UpperCamelCase_) , self.offset - 1) ] if len(set(UpperCamelCase_)) != len(UpperCamelCase_): raise ValueError( "Please make sure that the provided additional_special_tokens do not contain an incorrectly" F" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.") __UpperCAmelCase : str = additional_special_tokens_extended else: __UpperCAmelCase : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"<unk_{i}>" for i in range(2 , self.offset)] super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , pad_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , mask_token_sent=UpperCamelCase_ , offset=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Optional[int] = vocab_file __UpperCAmelCase : List[str] = False if not self.vocab_file else True def a_ ( self : Union[str, Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : int = set(self.all_special_ids) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens) + 3)): raise ValueError( "There should be 3 special tokens: mask_token, pad_token, and eos_token +" F" {len(self.additional_special_tokens)} additional_special_tokens, but got {all_special_ids}") return [1 if x in all_special_ids else 0 for x in seq] def a_ ( self : Union[str, Any] , UpperCamelCase_ : List , UpperCamelCase_ : Optional[List] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return self._special_token_mask(UpperCamelCase_) elif token_ids_a is None: return self._special_token_mask(UpperCamelCase_) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a) + [1] def a_ ( self : int , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : List[Any]=None): """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a_ ( self : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : List[str] = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_): copyfile(self.vocab_file , UpperCamelCase_) return (out_vocab_file,)
77
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class a__ : def __init__( self : List[str] , UpperCamelCase_ : List[str] , UpperCamelCase_ : int=2 , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : int=False , UpperCamelCase_ : List[str]=10 , UpperCamelCase_ : Tuple=3 , UpperCamelCase_ : int=32 * 4 , UpperCamelCase_ : Dict=32 * 6 , UpperCamelCase_ : Optional[int]=4 , UpperCamelCase_ : str=32 , ): """simple docstring""" __UpperCAmelCase : List[str] = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : List[str] = is_training __UpperCAmelCase : Optional[Any] = use_auxiliary_loss __UpperCAmelCase : Dict = num_queries __UpperCAmelCase : str = num_channels __UpperCAmelCase : Union[str, Any] = min_size __UpperCAmelCase : List[Any] = max_size __UpperCAmelCase : Any = num_labels __UpperCAmelCase : Optional[Any] = mask_feature_size def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size]).to( UpperCamelCase_) __UpperCAmelCase : str = torch.ones([self.batch_size, self.min_size, self.max_size] , device=UpperCamelCase_) __UpperCAmelCase : int = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=UpperCamelCase_) > 0.5 ).float() __UpperCAmelCase : Optional[int] = (torch.rand((self.batch_size, self.num_labels) , device=UpperCamelCase_) > 0.5).long() __UpperCAmelCase : Optional[Any] = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def a_ ( self : List[str]): """simple docstring""" return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=128 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Dict = self.prepare_config_and_inputs() __UpperCAmelCase : List[str] = {"pixel_values": pixel_values, "pixel_mask": pixel_mask} return config, inputs_dict def a_ ( self : List[Any] , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : str): """simple docstring""" __UpperCAmelCase : List[str] = output.encoder_hidden_states __UpperCAmelCase : Tuple = output.pixel_decoder_hidden_states __UpperCAmelCase : Tuple = output.transformer_decoder_hidden_states self.parent.assertTrue(len(UpperCamelCase_) , len(config.backbone_config.depths)) self.parent.assertTrue(len(UpperCamelCase_) , len(config.backbone_config.depths)) self.parent.assertTrue(len(UpperCamelCase_) , config.decoder_config.decoder_layers) def a_ ( self : Dict , UpperCamelCase_ : List[str] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : Dict , UpperCamelCase_ : Dict=False): """simple docstring""" with torch.no_grad(): __UpperCAmelCase : List[str] = MaskFormerModel(config=UpperCamelCase_) model.to(UpperCamelCase_) model.eval() __UpperCAmelCase : Union[str, Any] = model(pixel_values=UpperCamelCase_ , pixel_mask=UpperCamelCase_) __UpperCAmelCase : Dict = model(UpperCamelCase_ , output_hidden_states=UpperCamelCase_) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(output.encoder_last_hidden_state is not None) if output_hidden_states: self.check_output_hidden_state(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : str , UpperCamelCase_ : Tuple , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : Optional[int] , UpperCamelCase_ : List[Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : int = MaskFormerForInstanceSegmentation(config=UpperCamelCase_) model.to(UpperCamelCase_) model.eval() def comm_check_on_output(UpperCamelCase_ : Dict): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None) self.parent.assertTrue(result.encoder_last_hidden_state is not None) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1)) with torch.no_grad(): __UpperCAmelCase : Dict = model(pixel_values=UpperCamelCase_ , pixel_mask=UpperCamelCase_) __UpperCAmelCase : int = model(UpperCamelCase_) comm_check_on_output(UpperCamelCase_) __UpperCAmelCase : Any = model( pixel_values=UpperCamelCase_ , pixel_mask=UpperCamelCase_ , mask_labels=UpperCamelCase_ , class_labels=UpperCamelCase_) comm_check_on_output(UpperCamelCase_) self.parent.assertTrue(result.loss is not None) self.parent.assertEqual(result.loss.shape , torch.Size([1])) @require_torch class a__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase_ = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () lowercase_ = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) lowercase_ = False lowercase_ = False lowercase_ = False lowercase_ = False def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Tuple = MaskFormerModelTester(self) __UpperCAmelCase : List[str] = ConfigTester(self , config_class=UpperCamelCase_ , has_text_modality=UpperCamelCase_) def a_ ( self : int): """simple docstring""" self.config_tester.run_common_tests() def a_ ( self : List[Any]): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCamelCase_ , **UpperCamelCase_ , output_hidden_states=UpperCamelCase_) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*UpperCamelCase_) @unittest.skip(reason="MaskFormer does not use inputs_embeds") def a_ ( self : List[str]): """simple docstring""" pass @unittest.skip(reason="MaskFormer does not have a get_input_embeddings method") def a_ ( self : Any): """simple docstring""" pass @unittest.skip(reason="MaskFormer is not a generative model") def a_ ( self : Dict): """simple docstring""" pass @unittest.skip(reason="MaskFormer does not use token embeddings") def a_ ( self : List[str]): """simple docstring""" pass @require_torch_multi_gpu @unittest.skip( reason="MaskFormer has some layers using `add_module` which doesn't work well with `nn.DataParallel`") def a_ ( self : Optional[int]): """simple docstring""" pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests.") def a_ ( self : Union[str, Any]): """simple docstring""" pass def a_ ( self : str): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : Any = model_class(UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : List[str] = [*signature.parameters.keys()] __UpperCAmelCase : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase_) @slow def a_ ( self : int): """simple docstring""" for model_name in ["facebook/maskformer-swin-small-coco"]: __UpperCAmelCase : List[Any] = MaskFormerModel.from_pretrained(UpperCamelCase_) self.assertIsNotNone(UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : List[Any] = (self.model_tester.min_size,) * 2 __UpperCAmelCase : int = { "pixel_values": torch.randn((2, 3, *size) , device=UpperCamelCase_), "mask_labels": torch.randn((2, 10, *size) , device=UpperCamelCase_), "class_labels": torch.zeros(2 , 10 , device=UpperCamelCase_).long(), } __UpperCAmelCase : Union[str, Any] = MaskFormerForInstanceSegmentation(MaskFormerConfig()).to(UpperCamelCase_) __UpperCAmelCase : Dict = model(**UpperCamelCase_) self.assertTrue(outputs.loss is not None) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(UpperCamelCase_ , **UpperCamelCase_ , output_hidden_states=UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : str = model_class(UpperCamelCase_).to(UpperCamelCase_) __UpperCAmelCase : List[Any] = model(**UpperCamelCase_ , output_attentions=UpperCamelCase_) self.assertTrue(outputs.attentions is not None) def a_ ( self : str): """simple docstring""" if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss __UpperCAmelCase : Optional[int] = self.all_model_classes[1] __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() __UpperCAmelCase : Union[str, Any] = model_class(UpperCamelCase_) model.to(UpperCamelCase_) model.train() __UpperCAmelCase : List[str] = model(UpperCamelCase_ , mask_labels=UpperCamelCase_ , class_labels=UpperCamelCase_).loss loss.backward() def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Any = self.all_model_classes[1] __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() __UpperCAmelCase : List[Any] = True __UpperCAmelCase : Optional[Any] = True __UpperCAmelCase : Any = model_class(UpperCamelCase_) model.to(UpperCamelCase_) model.train() __UpperCAmelCase : Optional[int] = model(UpperCamelCase_ , mask_labels=UpperCamelCase_ , class_labels=UpperCamelCase_) __UpperCAmelCase : str = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() __UpperCAmelCase : Union[str, Any] = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't __UpperCAmelCase : Any = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() __UpperCAmelCase : Optional[int] = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=UpperCamelCase_) self.assertIsNotNone(encoder_hidden_states.grad) self.assertIsNotNone(pixel_decoder_hidden_states.grad) self.assertIsNotNone(transformer_decoder_hidden_states.grad) self.assertIsNotNone(attentions.grad) A = 1E-4 def _UpperCamelCase ( ) -> Any: """simple docstring""" __UpperCAmelCase : Union[str, Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @slow class a__ ( unittest.TestCase ): @cached_property def a_ ( self : Optional[Any]): """simple docstring""" return ( MaskFormerImageProcessor.from_pretrained("facebook/maskformer-swin-small-coco") if is_vision_available() else None ) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = MaskFormerModel.from_pretrained("facebook/maskformer-swin-small-coco").to(UpperCamelCase_) __UpperCAmelCase : Tuple = self.default_image_processor __UpperCAmelCase : Union[str, Any] = prepare_img() __UpperCAmelCase : Dict = image_processor(UpperCamelCase_ , return_tensors="pt").to(UpperCamelCase_) __UpperCAmelCase : str = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(UpperCamelCase_ , (1, 3, 800, 1088)) with torch.no_grad(): __UpperCAmelCase : Optional[int] = model(**UpperCamelCase_) __UpperCAmelCase : Tuple = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]]).to(UpperCamelCase_) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , UpperCamelCase_ , atol=UpperCamelCase_)) __UpperCAmelCase : List[Any] = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]]).to(UpperCamelCase_) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , UpperCamelCase_ , atol=UpperCamelCase_)) __UpperCAmelCase : List[Any] = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]]).to(UpperCamelCase_) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , UpperCamelCase_ , atol=UpperCamelCase_)) def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Union[str, Any] = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco") .to(UpperCamelCase_) .eval() ) __UpperCAmelCase : Optional[Any] = self.default_image_processor __UpperCAmelCase : Union[str, Any] = prepare_img() __UpperCAmelCase : Tuple = image_processor(UpperCamelCase_ , return_tensors="pt").to(UpperCamelCase_) __UpperCAmelCase : Dict = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(UpperCamelCase_ , (1, 3, 800, 1088)) with torch.no_grad(): __UpperCAmelCase : Optional[Any] = model(**UpperCamelCase_) # masks_queries_logits __UpperCAmelCase : Any = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) __UpperCAmelCase : str = [ [-1.3737124, -1.7724937, -1.9364233], [-1.5977281, -1.9867939, -2.1523695], [-1.5795398, -1.9269832, -2.093942], ] __UpperCAmelCase : List[str] = torch.tensor(UpperCamelCase_).to(UpperCamelCase_) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCamelCase_ , atol=UpperCamelCase_)) # class_queries_logits __UpperCAmelCase : Tuple = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1)) __UpperCAmelCase : Dict = torch.tensor( [ [1.6_512e00, -5.2_572e00, -3.3_519e00], [3.6_169e-02, -5.9_025e00, -2.9_313e00], [1.0_766e-04, -7.7_630e00, -5.1_263e00], ]).to(UpperCamelCase_) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCamelCase_ , atol=UpperCamelCase_)) def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Any = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-resnet101-coco-stuff") .to(UpperCamelCase_) .eval() ) __UpperCAmelCase : Dict = self.default_image_processor __UpperCAmelCase : List[str] = prepare_img() __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").to(UpperCamelCase_) __UpperCAmelCase : str = inputs["pixel_values"].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0) # check size self.assertEqual(UpperCamelCase_ , (1, 3, 800, 1088)) with torch.no_grad(): __UpperCAmelCase : str = model(**UpperCamelCase_) # masks_queries_logits __UpperCAmelCase : Optional[Any] = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) __UpperCAmelCase : Any = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] __UpperCAmelCase : Any = torch.tensor(UpperCamelCase_).to(UpperCamelCase_) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , UpperCamelCase_ , atol=UpperCamelCase_)) # class_queries_logits __UpperCAmelCase : Optional[Any] = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1)) __UpperCAmelCase : str = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]]).to(UpperCamelCase_) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCamelCase_ , atol=UpperCamelCase_)) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[int] = ( MaskFormerForInstanceSegmentation.from_pretrained("facebook/maskformer-swin-small-coco") .to(UpperCamelCase_) .eval() ) __UpperCAmelCase : int = self.default_image_processor __UpperCAmelCase : List[Any] = image_processor( [np.zeros((3, 800, 1333)), np.zeros((3, 800, 1333))] , segmentation_maps=[np.zeros((384, 384)).astype(np.floataa), np.zeros((384, 384)).astype(np.floataa)] , return_tensors="pt" , ) __UpperCAmelCase : Tuple = inputs["pixel_values"].to(UpperCamelCase_) __UpperCAmelCase : int = [el.to(UpperCamelCase_) for el in inputs["mask_labels"]] __UpperCAmelCase : int = [el.to(UpperCamelCase_) for el in inputs["class_labels"]] with torch.no_grad(): __UpperCAmelCase : List[Any] = model(**UpperCamelCase_) self.assertTrue(outputs.loss is not None)
77
"""simple docstring""" import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> List[str]: """simple docstring""" # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file __UpperCAmelCase : Optional[Any] = TapasConfig.from_json_file(UpperCamelCase ) # set absolute/relative position embeddings parameter __UpperCAmelCase : Optional[Any] = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": __UpperCAmelCase : List[str] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WTQ": # run_task_main.py hparams __UpperCAmelCase : Tuple = 4 __UpperCAmelCase : Any = True # hparam_utils.py hparams __UpperCAmelCase : Union[str, Any] = 0.664694 __UpperCAmelCase : Union[str, Any] = 0.207951 __UpperCAmelCase : int = 0.121194 __UpperCAmelCase : Optional[int] = True __UpperCAmelCase : List[str] = True __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : List[str] = 0.0352513 __UpperCAmelCase : Optional[int] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams __UpperCAmelCase : int = 4 __UpperCAmelCase : Optional[int] = False # hparam_utils.py hparams __UpperCAmelCase : int = 36.4519 __UpperCAmelCase : str = 0.903421 __UpperCAmelCase : Dict = 222.088 __UpperCAmelCase : Dict = True __UpperCAmelCase : Union[str, Any] = True __UpperCAmelCase : Tuple = True __UpperCAmelCase : Any = 0.763141 __UpperCAmelCase : Optional[Any] = TapasForQuestionAnswering(config=UpperCamelCase ) elif task == "TABFACT": __UpperCAmelCase : Union[str, Any] = TapasForSequenceClassification(config=UpperCamelCase ) elif task == "MLM": __UpperCAmelCase : Tuple = TapasForMaskedLM(config=UpperCamelCase ) elif task == "INTERMEDIATE_PRETRAINING": __UpperCAmelCase : List[str] = TapasModel(config=UpperCamelCase ) else: raise ValueError(f"Task {task} not supported." ) print(f"Building PyTorch model from configuration: {config}" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # Save pytorch-model (weights and configuration) print(f"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(UpperCamelCase ) # Save tokenizer files print(f"Save tokenizer files to {pytorch_dump_path}" ) __UpperCAmelCase : str = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + "vocab.txt" , model_max_length=512 ) tokenizer.save_pretrained(UpperCamelCase ) print("Used relative position embeddings:" , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""SQA""", type=str, help="""Model task for which to convert a checkpoint. Defaults to SQA.""" ) parser.add_argument( """--reset_position_index_per_cell""", default=False, action="""store_true""", help="""Whether to use relative position embeddings or not. Defaults to True.""", ) parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--tapas_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained TAPAS model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
77
1
"""simple docstring""" import re def _UpperCamelCase ( UpperCamelCase ) -> list: """simple docstring""" return [char.split() for char in re.split(R"[^ a-z A-Z 0-9 \s]" , str_ )] def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" __UpperCAmelCase : Optional[int] = split_input(str_ ) return "".join( ["".join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" try: __UpperCAmelCase : Tuple = split_input(UpperCamelCase ) if upper: __UpperCAmelCase : Any = "".join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: __UpperCAmelCase : int = "".join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" return to_simple_case(UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase ) -> str: """simple docstring""" try: __UpperCAmelCase : Optional[Any] = to_simple_case(UpperCamelCase ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" return to_complex_case(UpperCamelCase , UpperCamelCase , "_" ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> str: """simple docstring""" return to_complex_case(UpperCamelCase , UpperCamelCase , "-" ) if __name__ == "__main__": __import__("""doctest""").testmod()
77
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = "cpu" , UpperCamelCase = None ) -> None: """simple docstring""" __UpperCAmelCase : Union[str, Any] = torch.load(UpperCamelCase , map_location=UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(UpperCamelCase , torch.Tensor ): raise TypeError("FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin" ) __UpperCAmelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __UpperCAmelCase : str = src_path torch.save(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
77
1
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import OwlViTImageProcessor, OwlViTProcessor @require_vision class a__ ( unittest.TestCase ): def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : str = tempfile.mkdtemp() # fmt: off __UpperCAmelCase : List[Any] = ["", "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on __UpperCAmelCase : Union[str, Any] = dict(zip(UpperCamelCase_ , range(len(UpperCamelCase_)))) __UpperCAmelCase : Dict = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] __UpperCAmelCase : str = {"unk_token": "<unk>"} __UpperCAmelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"]) __UpperCAmelCase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"]) with open(self.vocab_file , "w" , encoding="utf-8") as fp: fp.write(json.dumps(UpperCamelCase_) + "\n") with open(self.merges_file , "w" , encoding="utf-8") as fp: fp.write("\n".join(UpperCamelCase_)) __UpperCAmelCase : str = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48145466, 0.4578275, 0.40821073], "image_std": [0.26862954, 0.26130258, 0.27577711], } __UpperCAmelCase : List[Any] = os.path.join(self.tmpdirname , UpperCamelCase_) with open(self.image_processor_file , "w" , encoding="utf-8") as fp: json.dump(UpperCamelCase_ , UpperCamelCase_) def a_ ( self : Any , **UpperCamelCase_ : Dict): """simple docstring""" return CLIPTokenizer.from_pretrained(self.tmpdirname , pad_token="!" , **UpperCamelCase_) def a_ ( self : Union[str, Any] , **UpperCamelCase_ : List[Any]): """simple docstring""" return CLIPTokenizerFast.from_pretrained(self.tmpdirname , pad_token="!" , **UpperCamelCase_) def a_ ( self : Tuple , **UpperCamelCase_ : Union[str, Any]): """simple docstring""" return OwlViTImageProcessor.from_pretrained(self.tmpdirname , **UpperCamelCase_) def a_ ( self : Union[str, Any]): """simple docstring""" shutil.rmtree(self.tmpdirname) def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Dict = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta)] __UpperCAmelCase : List[str] = [Image.fromarray(np.moveaxis(UpperCamelCase_ , 0 , -1)) for x in image_inputs] return image_inputs def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : int = self.get_tokenizer() __UpperCAmelCase : int = self.get_rust_tokenizer() __UpperCAmelCase : int = self.get_image_processor() __UpperCAmelCase : Tuple = OwlViTProcessor(tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_) processor_slow.save_pretrained(self.tmpdirname) __UpperCAmelCase : Union[str, Any] = OwlViTProcessor.from_pretrained(self.tmpdirname , use_fast=UpperCamelCase_) __UpperCAmelCase : List[Any] = OwlViTProcessor(tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_) processor_fast.save_pretrained(self.tmpdirname) __UpperCAmelCase : Optional[Any] = OwlViTProcessor.from_pretrained(self.tmpdirname) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab()) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab()) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab()) self.assertIsInstance(processor_slow.tokenizer , UpperCamelCase_) self.assertIsInstance(processor_fast.tokenizer , UpperCamelCase_) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string()) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string()) self.assertIsInstance(processor_slow.image_processor , UpperCamelCase_) self.assertIsInstance(processor_fast.image_processor , UpperCamelCase_) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : List[Any] = OwlViTProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) __UpperCAmelCase : Dict = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)") __UpperCAmelCase : str = self.get_image_processor(do_normalize=UpperCamelCase_) __UpperCAmelCase : Any = OwlViTProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=UpperCamelCase_) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.tokenizer , UpperCamelCase_) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = self.get_image_processor() __UpperCAmelCase : List[str] = self.get_tokenizer() __UpperCAmelCase : Optional[int] = OwlViTProcessor(tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_) __UpperCAmelCase : str = self.prepare_image_inputs() __UpperCAmelCase : Union[str, Any] = image_processor(UpperCamelCase_ , return_tensors="np") __UpperCAmelCase : List[str] = processor(images=UpperCamelCase_ , return_tensors="np") for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2) def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[int] = self.get_image_processor() __UpperCAmelCase : Union[str, Any] = self.get_tokenizer() __UpperCAmelCase : str = OwlViTProcessor(tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_) __UpperCAmelCase : Tuple = "lower newer" __UpperCAmelCase : List[Any] = processor(text=UpperCamelCase_ , return_tensors="np") __UpperCAmelCase : Any = tokenizer(UpperCamelCase_ , return_tensors="np") for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key][0].tolist() , encoded_processor[key][0].tolist()) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : List[Any] = self.get_image_processor() __UpperCAmelCase : int = self.get_tokenizer() __UpperCAmelCase : Tuple = OwlViTProcessor(tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_) __UpperCAmelCase : Union[str, Any] = "lower newer" __UpperCAmelCase : str = self.prepare_image_inputs() __UpperCAmelCase : Any = processor(text=UpperCamelCase_ , images=UpperCamelCase_) self.assertListEqual(list(inputs.keys()) , ["input_ids", "attention_mask", "pixel_values"]) # test if it raises when no input is passed with pytest.raises(UpperCamelCase_): processor() def a_ ( self : str): """simple docstring""" __UpperCAmelCase : Dict = "google/owlvit-base-patch32" __UpperCAmelCase : str = OwlViTProcessor.from_pretrained(UpperCamelCase_) __UpperCAmelCase : int = ["cat", "nasa badge"] __UpperCAmelCase : Dict = processor(text=UpperCamelCase_) __UpperCAmelCase : List[Any] = 16 self.assertListEqual(list(inputs.keys()) , ["input_ids", "attention_mask"]) self.assertEqual(inputs["input_ids"].shape , (2, seq_length)) # test if it raises when no input is passed with pytest.raises(UpperCamelCase_): processor() def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[Any] = "google/owlvit-base-patch32" __UpperCAmelCase : Tuple = OwlViTProcessor.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[int] = [["cat", "nasa badge"], ["person"]] __UpperCAmelCase : List[str] = processor(text=UpperCamelCase_) __UpperCAmelCase : List[str] = 16 __UpperCAmelCase : str = len(UpperCamelCase_) __UpperCAmelCase : List[str] = max([len(UpperCamelCase_) for texts in input_texts]) self.assertListEqual(list(inputs.keys()) , ["input_ids", "attention_mask"]) self.assertEqual(inputs["input_ids"].shape , (batch_size * num_max_text_queries, seq_length)) # test if it raises when no input is passed with pytest.raises(UpperCamelCase_): processor() def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Tuple = "google/owlvit-base-patch32" __UpperCAmelCase : Dict = OwlViTProcessor.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Any = ["cat", "nasa badge"] __UpperCAmelCase : Optional[int] = processor(text=UpperCamelCase_) __UpperCAmelCase : Optional[int] = 16 __UpperCAmelCase : Optional[Any] = inputs["input_ids"] __UpperCAmelCase : Any = [ [49406, 2368, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [49406, 6841, 11301, 49407, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] self.assertListEqual(list(inputs.keys()) , ["input_ids", "attention_mask"]) self.assertEqual(inputs["input_ids"].shape , (2, seq_length)) self.assertListEqual(list(input_ids[0]) , predicted_ids[0]) self.assertListEqual(list(input_ids[1]) , predicted_ids[1]) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Any = self.get_image_processor() __UpperCAmelCase : Any = self.get_tokenizer() __UpperCAmelCase : Tuple = OwlViTProcessor(tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_) __UpperCAmelCase : Any = self.prepare_image_inputs() __UpperCAmelCase : List[Any] = self.prepare_image_inputs() __UpperCAmelCase : Optional[Any] = processor(images=UpperCamelCase_ , query_images=UpperCamelCase_) self.assertListEqual(list(inputs.keys()) , ["query_pixel_values", "pixel_values"]) # test if it raises when no input is passed with pytest.raises(UpperCamelCase_): processor() def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.get_image_processor() __UpperCAmelCase : str = self.get_tokenizer() __UpperCAmelCase : str = OwlViTProcessor(tokenizer=UpperCamelCase_ , image_processor=UpperCamelCase_) __UpperCAmelCase : Tuple = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __UpperCAmelCase : List[str] = processor.batch_decode(UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer.batch_decode(UpperCamelCase_) self.assertListEqual(UpperCamelCase_ , UpperCamelCase_)
77
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": A = pd.read_csv("""sample_data.csv""", header=None) A = df.shape[:1][0] # If you're using some other dataset input the target column A = df.iloc[:, 1:2] A = actual_data.values.reshape(len_data, 1) A = MinMaxScaler().fit_transform(actual_data) A = 10 A = 5 A = 20 A = len_data - periods * look_back A = actual_data[:division] A = actual_data[division - look_back :] A , A = [], [] A , A = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) A = np.array(train_x) A = np.array(test_x) A = np.array([list(i.ravel()) for i in train_y]) A = np.array([list(i.ravel()) for i in test_y]) A = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") A = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) A = model.predict(x_test)
77
1
"""simple docstring""" import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) A = """▁""" A = {"""vocab_file""": """vocab.txt""", """sentencepiece_model_ckpt""": """sentencepiece.bpe.model"""} A = { """sentencepiece_model_file""": """sentencepiece.bpe.model""", """vocab_file""": """vocab.txt""", } A = { """vocab_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt""", }, """sentencepiece_model_file""": { """ernie-m-base""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", """ernie-m-large""": """https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model""", }, } A = { """ernie-m-base""": 514, """ernie-m-large""": 514, } A = { """ernie-m-base""": {"""do_lower_case""": False}, """ernie-m-large""": {"""do_lower_case""": False}, } class a__ ( __magic_name__ ): lowercase_ = ["input_ids"] lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_INIT_CONFIGURATION lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = RESOURCE_FILES_NAMES def __init__( self : Dict , UpperCamelCase_ : Tuple , UpperCamelCase_ : List[str]=None , UpperCamelCase_ : List[str]=False , UpperCamelCase_ : List[str]="utf8" , UpperCamelCase_ : Tuple="[UNK]" , UpperCamelCase_ : Optional[int]="[SEP]" , UpperCamelCase_ : str="[PAD]" , UpperCamelCase_ : int="[CLS]" , UpperCamelCase_ : str="[MASK]" , UpperCamelCase_ : Optional[Dict[str, Any]] = None , **UpperCamelCase_ : int , ): """simple docstring""" __UpperCAmelCase : Any = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , vocab_file=UpperCamelCase_ , encoding=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) __UpperCAmelCase : Tuple = do_lower_case __UpperCAmelCase : Any = sentencepiece_model_ckpt __UpperCAmelCase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(UpperCamelCase_) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: __UpperCAmelCase : List[Any] = self.load_vocab(filepath=UpperCamelCase_) else: __UpperCAmelCase : str = {self.sp_model.id_to_piece(UpperCamelCase_): id for id in range(self.sp_model.get_piece_size())} __UpperCAmelCase : Optional[int] = {v: k for k, v in self.vocab.items()} def a_ ( self : Tuple , UpperCamelCase_ : Union[str, Any]): """simple docstring""" if text is None: return None __UpperCAmelCase : str = self.tokenize(UpperCamelCase_) __UpperCAmelCase , __UpperCAmelCase : List[Any] = "", [] for i, ch in enumerate(UpperCamelCase_): if ch in self.SP_CHAR_MAPPING: __UpperCAmelCase : int = self.SP_CHAR_MAPPING.get(UpperCamelCase_) else: __UpperCAmelCase : List[str] = unicodedata.normalize("NFKC" , UpperCamelCase_) if self.is_whitespace(UpperCamelCase_): continue normalized_text += ch char_mapping.extend([i] * len(UpperCamelCase_)) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : List[str] = normalized_text, [], 0 if self.do_lower_case: __UpperCAmelCase : List[Any] = text.lower() for token in split_tokens: if token[:1] == "▁": __UpperCAmelCase : Tuple = token[1:] __UpperCAmelCase : Tuple = text[offset:].index(UpperCamelCase_) + offset __UpperCAmelCase : Union[str, Any] = start + len(UpperCamelCase_) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1)) __UpperCAmelCase : Dict = end return token_mapping @property def a_ ( self : Union[str, Any]): """simple docstring""" return len(self.vocab) def a_ ( self : Any): """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder) def __getstate__( self : int): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.__dict__.copy() __UpperCAmelCase : Tuple = None return state def __setstate__( self : str , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : int = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): __UpperCAmelCase : List[str] = {} __UpperCAmelCase : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.sentencepiece_model_ckpt) def a_ ( self : Any , UpperCamelCase_ : int): """simple docstring""" return "".join((self.SP_CHAR_MAPPING.get(UpperCamelCase_ , UpperCamelCase_) for c in text)) def a_ ( self : Dict , UpperCamelCase_ : str , UpperCamelCase_ : Dict=False , UpperCamelCase_ : Dict=64 , UpperCamelCase_ : Optional[Any]=0.1): """simple docstring""" if self.sp_model_kwargs.get("enable_sampling") is True: __UpperCAmelCase : Optional[int] = True if self.sp_model_kwargs.get("alpha") is not None: __UpperCAmelCase : List[str] = self.sp_model_kwargs.get("alpha") if self.sp_model_kwargs.get("nbest_size") is not None: __UpperCAmelCase : Optional[Any] = self.sp_model_kwargs.get("nbest_size") if not enable_sampling: __UpperCAmelCase : List[str] = self.sp_model.EncodeAsPieces(UpperCamelCase_) else: __UpperCAmelCase : List[str] = self.sp_model.SampleEncodeAsPieces(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Dict = [] for pi, piece in enumerate(UpperCamelCase_): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(UpperCamelCase_) and pi != 0: new_pieces.append(UpperCamelCase_) continue else: continue __UpperCAmelCase : Optional[Any] = 0 for i, chunk in enumerate(UpperCamelCase_): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(UpperCamelCase_) or self.is_punct(UpperCamelCase_): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i]) new_pieces.append(UpperCamelCase_) __UpperCAmelCase : int = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i]) __UpperCAmelCase : Tuple = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i]) __UpperCAmelCase : Dict = i if len(UpperCamelCase_) > lst_i: new_pieces.append(piece[lst_i:]) return new_pieces def a_ ( self : Tuple , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : int = "".join(UpperCamelCase_).replace(UpperCamelCase_ , " ").strip() return out_string def a_ ( self : Optional[Any] , UpperCamelCase_ : List[Any]): """simple docstring""" __UpperCAmelCase : Dict = self.convert_ids_to_tokens(UpperCamelCase_) __UpperCAmelCase : str = "".join(UpperCamelCase_).replace(UpperCamelCase_ , " ").strip() return out_string def a_ ( self : Union[str, Any] , UpperCamelCase_ : int): """simple docstring""" return self.vocab.get(UpperCamelCase_ , self.vocab.get(self.unk_token)) def a_ ( self : Optional[int] , UpperCamelCase_ : Optional[int]): """simple docstring""" return self.reverse_vocab.get(UpperCamelCase_ , self.unk_token) def a_ ( self : Any , UpperCamelCase_ : List[Any] , UpperCamelCase_ : List[str]=None): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __UpperCAmelCase : str = [self.cls_token_id] __UpperCAmelCase : Optional[Any] = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def a_ ( self : List[Any] , UpperCamelCase_ : Optional[Any] , UpperCamelCase_ : str=None): """simple docstring""" if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def a_ ( self : str , UpperCamelCase_ : str , UpperCamelCase_ : int=None , UpperCamelCase_ : str=False): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model.") return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(UpperCamelCase_)) + [1, 1] + ([0] * len(UpperCamelCase_)) + [1] return [1] + ([0] * len(UpperCamelCase_)) + [1] def a_ ( self : Tuple , UpperCamelCase_ : List[int] , UpperCamelCase_ : Optional[List[int]] = None): """simple docstring""" if token_ids_a is None: # [CLS] X [SEP] return (len(UpperCamelCase_) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(UpperCamelCase_) + 1) + [1] * (len(UpperCamelCase_) + 3) def a_ ( self : Optional[int] , UpperCamelCase_ : Optional[int]): """simple docstring""" if "\u4e00" <= char <= "\u9fff": return True return False def a_ ( self : List[str] , UpperCamelCase_ : Optional[Any]): """simple docstring""" if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def a_ ( self : List[Any] , UpperCamelCase_ : Any): """simple docstring""" if char in ",;:.?!~,;:。?!《》【】": return True return False def a_ ( self : Optional[int] , UpperCamelCase_ : List[Any]): """simple docstring""" if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(UpperCamelCase_) == 1: __UpperCAmelCase : Dict = unicodedata.category(UpperCamelCase_) if cat == "Zs": return True return False def a_ ( self : Union[str, Any] , UpperCamelCase_ : int): """simple docstring""" __UpperCAmelCase : Optional[int] = {} with io.open(UpperCamelCase_ , "r" , encoding="utf-8") as f: for index, line in enumerate(UpperCamelCase_): __UpperCAmelCase : Optional[int] = line.rstrip("\n") __UpperCAmelCase : int = int(UpperCamelCase_) return token_to_idx def a_ ( self : List[str] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" __UpperCAmelCase : Optional[Any] = 0 if os.path.isdir(UpperCamelCase_): __UpperCAmelCase : Tuple = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) else: __UpperCAmelCase : int = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(UpperCamelCase_ , "w" , encoding="utf-8") as writer: for token, token_index in sorted(self.vocab.items() , key=lambda UpperCamelCase_: kv[1]): if index != token_index: logger.warning( F"Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive." " Please check that the vocabulary is not corrupted!") __UpperCAmelCase : List[Any] = token_index writer.write(token + "\n") index += 1 __UpperCAmelCase : Optional[Any] = os.path.join(UpperCamelCase_ , "sentencepiece.bpe.model") with open(UpperCamelCase_ , "wb") as fi: __UpperCAmelCase : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_) return (vocab_file,)
77
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartTokenizer, MBartTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin A = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right A = 250_004 A = 250_020 @require_sentencepiece @require_tokenizers class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = MBartTokenizer lowercase_ = MBartTokenizerFast lowercase_ = True lowercase_ = True def a_ ( self : str): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase : Any = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) tokenizer.save_pretrained(self.tmpdirname) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Dict = MBartTokenizer(UpperCamelCase_ , keep_accents=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer.tokenize("This is a test") self.assertListEqual(UpperCamelCase_ , ["▁This", "▁is", "▁a", "▁t", "est"]) self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCamelCase_) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __UpperCAmelCase : List[Any] = tokenizer.tokenize("I was born in 92000, and this is falsé.") self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) __UpperCAmelCase : Any = tokenizer.convert_tokens_to_ids(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) __UpperCAmelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(UpperCamelCase_) self.assertListEqual( UpperCamelCase_ , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) def a_ ( self : Dict): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __UpperCAmelCase : Dict = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})"): __UpperCAmelCase : List[str] = self.rust_tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer_class.from_pretrained(UpperCamelCase_ , **UpperCamelCase_) __UpperCAmelCase : int = tempfile.mkdtemp() __UpperCAmelCase : Optional[int] = tokenizer_r.save_pretrained(UpperCamelCase_) __UpperCAmelCase : Any = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) __UpperCAmelCase : Any = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f) self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Tuple = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=True __UpperCAmelCase : Optional[int] = tempfile.mkdtemp() __UpperCAmelCase : Dict = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : int = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it save with the same files self.assertSequenceEqual(UpperCamelCase_ , UpperCamelCase_) # Checks everything loads correctly in the same way __UpperCAmelCase : int = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : Optional[Any] = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) # Save tokenizer rust, legacy_format=False __UpperCAmelCase : Tuple = tempfile.mkdtemp() __UpperCAmelCase : int = tokenizer_r.save_pretrained(UpperCamelCase_ , legacy_format=UpperCamelCase_) __UpperCAmelCase : Optional[int] = tokenizer_p.save_pretrained(UpperCamelCase_) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files)) # Checks everything loads correctly in the same way __UpperCAmelCase : Optional[Any] = tokenizer_r.from_pretrained(UpperCamelCase_) __UpperCAmelCase : str = tokenizer_p.from_pretrained(UpperCamelCase_) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(UpperCamelCase_ , UpperCamelCase_)) shutil.rmtree(UpperCamelCase_) @require_torch @require_sentencepiece @require_tokenizers class a__ ( unittest.TestCase ): lowercase_ = "facebook/mbart-large-en-ro" lowercase_ = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] lowercase_ = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] lowercase_ = [8_2_7_4, 1_2_7_8_7_3, 2_5_9_1_6, 7, 8_6_2_2, 2_0_7_1, 4_3_8, 6_7_4_8_5, 5_3, 1_8_7_8_9_5, 2_3, 5_1_7_1_2, 2, EN_CODE] @classmethod def a_ ( cls : int): """simple docstring""" __UpperCAmelCase : MBartTokenizer = MBartTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en_XX" , tgt_lang="ro_RO") __UpperCAmelCase : Union[str, Any] = 1 return cls def a_ ( self : List[Any]): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ar_AR"] , 250001) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["en_EN"] , 250004) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ro_RO"] , 250020) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.tokenizer.batch_encode_plus(self.src_text).input_ids[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) def a_ ( self : Optional[int]): """simple docstring""" self.assertIn(UpperCamelCase_ , self.tokenizer.all_special_ids) __UpperCAmelCase : Union[str, Any] = [RO_CODE, 884, 9019, 96, 9, 916, 86792, 36, 18743, 15596, 5, 2] __UpperCAmelCase : Optional[Any] = self.tokenizer.decode(UpperCamelCase_ , skip_special_tokens=UpperCamelCase_) __UpperCAmelCase : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=UpperCamelCase_) self.assertEqual(UpperCamelCase_ , UpperCamelCase_) self.assertNotIn(self.tokenizer.eos_token , UpperCamelCase_) def a_ ( self : int): """simple docstring""" __UpperCAmelCase : Optional[Any] = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , UpperCamelCase_) __UpperCAmelCase : Tuple = 10 __UpperCAmelCase : List[Any] = self.tokenizer(UpperCamelCase_ , max_length=UpperCamelCase_ , truncation=UpperCamelCase_).input_ids[0] self.assertEqual(ids[-2] , 2) self.assertEqual(ids[-1] , UpperCamelCase_) self.assertEqual(len(UpperCamelCase_) , UpperCamelCase_) def a_ ( self : Any): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"]) , [250026, 250001]) def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : List[str] = tempfile.mkdtemp() __UpperCAmelCase : Union[str, Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(UpperCamelCase_) __UpperCAmelCase : List[Any] = MBartTokenizer.from_pretrained(UpperCamelCase_) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , UpperCamelCase_) @require_torch def a_ ( self : Optional[Any]): """simple docstring""" __UpperCAmelCase : Union[str, Any] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , return_tensors="pt") __UpperCAmelCase : Dict = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][-2:].tolist() == [2, EN_CODE] assert batch.decoder_input_ids[1][0].tolist() == RO_CODE assert batch.decoder_input_ids[1][-1] == 2 assert batch.labels[1][-2:].tolist() == [2, RO_CODE] @require_torch def a_ ( self : Optional[int]): """simple docstring""" __UpperCAmelCase : Dict = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=len(self.expected_src_tokens) , return_tensors="pt" , ) __UpperCAmelCase : Tuple = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_) self.assertEqual((2, 14) , batch.input_ids.shape) self.assertEqual((2, 14) , batch.attention_mask.shape) __UpperCAmelCase : List[str] = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , UpperCamelCase_) self.assertEqual(2 , batch.decoder_input_ids[0, -1]) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , []) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, EN_CODE]) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : List[str] = self.tokenizer(self.src_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=3 , return_tensors="pt") __UpperCAmelCase : Any = self.tokenizer( text_target=self.tgt_text , padding=UpperCamelCase_ , truncation=UpperCamelCase_ , max_length=10 , return_tensors="pt") __UpperCAmelCase : int = targets["input_ids"] __UpperCAmelCase : Any = shift_tokens_right(UpperCamelCase_ , self.tokenizer.pad_token_id) self.assertEqual(batch.input_ids.shape[1] , 3) self.assertEqual(batch.decoder_input_ids.shape[1] , 10) @require_torch def a_ ( self : int): """simple docstring""" __UpperCAmelCase : int = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="ar_AR") self.assertEqual( nested_simplify(UpperCamelCase_) , { # A, test, EOS, en_XX "input_ids": [[62, 3034, 2, 250004]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 250001, } , )
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> bool: """simple docstring""" if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("check_bouncy() accepts only integer arguments" ) __UpperCAmelCase : Optional[int] = str(UpperCamelCase ) __UpperCAmelCase : List[Any] = "".join(sorted(UpperCamelCase ) ) return sorted_str_n != str_n and sorted_str_n[::-1] != str_n def _UpperCamelCase ( UpperCamelCase = 99 ) -> int: """simple docstring""" if not 0 < percent < 100: raise ValueError("solution() only accepts values from 0 to 100" ) __UpperCAmelCase : List[Any] = 0 __UpperCAmelCase : List[Any] = 1 while True: if check_bouncy(UpperCamelCase ): bouncy_num += 1 if (bouncy_num / num) * 100 >= percent: return num num += 1 if __name__ == "__main__": from doctest import testmod testmod() print(f'''{solution(99)}''')
77
"""simple docstring""" from typing import Any class a__ : def __init__( self : List[str] , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : str = data __UpperCAmelCase : Optional[Any] = None class a__ : def __init__( self : Any): """simple docstring""" __UpperCAmelCase : Optional[int] = None def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.head while temp is not None: print(temp.data , end=" ") __UpperCAmelCase : Tuple = temp.next print() def a_ ( self : int , UpperCamelCase_ : Any): """simple docstring""" __UpperCAmelCase : List[str] = Node(UpperCamelCase_) __UpperCAmelCase : str = self.head __UpperCAmelCase : Optional[int] = new_node def a_ ( self : str , UpperCamelCase_ : List[Any] , UpperCamelCase_ : str): """simple docstring""" if node_data_a == node_data_a: return else: __UpperCAmelCase : int = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Tuple = node_a.next __UpperCAmelCase : List[Any] = self.head while node_a is not None and node_a.data != node_data_a: __UpperCAmelCase : Optional[Any] = node_a.next if node_a is None or node_a is None: return __UpperCAmelCase , __UpperCAmelCase : Any = node_a.data, node_a.data if __name__ == "__main__": A = LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
77
1
"""simple docstring""" import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class a__ ( unittest.TestCase ): lowercase_ = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def a_ ( self : Tuple , UpperCamelCase_ : List[Any] , UpperCamelCase_ : str , UpperCamelCase_ : List[str]): """simple docstring""" __UpperCAmelCase : Any = hf_hub_download( repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset") __UpperCAmelCase : List[Any] = VideoClassificationPipeline(model=UpperCamelCase_ , image_processor=UpperCamelCase_ , top_k=2) __UpperCAmelCase : Optional[int] = [ example_video_filepath, "https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4", ] return video_classifier, examples def a_ ( self : Union[str, Any] , UpperCamelCase_ : Any , UpperCamelCase_ : Union[str, Any]): """simple docstring""" for example in examples: __UpperCAmelCase : List[str] = video_classifier(UpperCamelCase_) self.assertEqual( UpperCamelCase_ , [ {"score": ANY(UpperCamelCase_), "label": ANY(UpperCamelCase_)}, {"score": ANY(UpperCamelCase_), "label": ANY(UpperCamelCase_)}, ] , ) @require_torch def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : str = "hf-internal-testing/tiny-random-VideoMAEForVideoClassification" __UpperCAmelCase : Tuple = VideoMAEFeatureExtractor( size={"shortest_edge": 10} , crop_size={"height": 10, "width": 10}) __UpperCAmelCase : str = pipeline( "video-classification" , model=UpperCamelCase_ , feature_extractor=UpperCamelCase_ , frame_sampling_rate=4) __UpperCAmelCase : Tuple = hf_hub_download(repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset") __UpperCAmelCase : List[str] = video_classifier(UpperCamelCase_ , top_k=2) self.assertEqual( nested_simplify(UpperCamelCase_ , decimals=4) , [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}] , ) __UpperCAmelCase : Union[str, Any] = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(UpperCamelCase_ , decimals=4) , [ [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}], [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}], ] , ) @require_tf def a_ ( self : str): """simple docstring""" pass
77
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore A = """ Human: <<task>> Assistant: """ A = """huggingface-tools/default-prompts""" A = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase="run" ) -> List[str]: """simple docstring""" if prompt_or_repo_id is None: __UpperCAmelCase : Optional[int] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , UpperCamelCase ) is not None: return prompt_or_repo_id __UpperCAmelCase : str = cached_file( UpperCamelCase , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name} ) with open(UpperCamelCase , "r" , encoding="utf-8" ) as f: return f.read()
77
1
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("multiplicative_persistence() only accepts integral values" ) if num < 0: raise ValueError("multiplicative_persistence() does not accept negative values" ) __UpperCAmelCase : List[Any] = 0 __UpperCAmelCase : str = str(UpperCamelCase ) while len(UpperCamelCase ) != 1: __UpperCAmelCase : Any = [int(UpperCamelCase ) for i in num_string] __UpperCAmelCase : List[str] = 1 for i in range(0 , len(UpperCamelCase ) ): total *= numbers[i] __UpperCAmelCase : Optional[int] = str(UpperCamelCase ) steps += 1 return steps def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if not isinstance(UpperCamelCase , UpperCamelCase ): raise ValueError("additive_persistence() only accepts integral values" ) if num < 0: raise ValueError("additive_persistence() does not accept negative values" ) __UpperCAmelCase : Dict = 0 __UpperCAmelCase : List[str] = str(UpperCamelCase ) while len(UpperCamelCase ) != 1: __UpperCAmelCase : Dict = [int(UpperCamelCase ) for i in num_string] __UpperCAmelCase : int = 0 for i in range(0 , len(UpperCamelCase ) ): total += numbers[i] __UpperCAmelCase : Optional[int] = str(UpperCamelCase ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available A = { """configuration_ernie""": ["""ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ErnieConfig""", """ErnieOnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST""", """ErnieForCausalLM""", """ErnieForMaskedLM""", """ErnieForMultipleChoice""", """ErnieForNextSentencePrediction""", """ErnieForPreTraining""", """ErnieForQuestionAnswering""", """ErnieForSequenceClassification""", """ErnieForTokenClassification""", """ErnieModel""", """ErniePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" from __future__ import annotations from math import gcd def _UpperCamelCase ( UpperCamelCase , UpperCamelCase = 2 , UpperCamelCase = 1 , UpperCamelCase = 3 , ) -> int | None: """simple docstring""" # A value less than 2 can cause an infinite loop in the algorithm. if num < 2: raise ValueError("The input value cannot be less than 2" ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> int: return (pow(UpperCamelCase , 2 ) + step) % modulus for _ in range(UpperCamelCase ): # These track the position within the cycle detection logic. __UpperCAmelCase : int = seed __UpperCAmelCase : Tuple = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. __UpperCAmelCase : Tuple = rand_fn(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : Tuple = rand_fn(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __UpperCAmelCase : str = rand_fn(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. __UpperCAmelCase : Union[str, Any] = gcd(hare - tortoise , UpperCamelCase ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. __UpperCAmelCase : List[str] = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse A = argparse.ArgumentParser() parser.add_argument( """num""", type=int, help="""The value to find a divisor of""", ) parser.add_argument( """--attempts""", type=int, default=3, help="""The number of attempts before giving up""", ) A = parser.parse_args() A = pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(f'''{args.num} is probably prime''') else: A = args.num // divisor print(f'''{args.num} = {divisor} * {quotient}''')
77
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class a__ ( nn.Module ): def __init__( self : Union[str, Any]): """simple docstring""" super().__init__() __UpperCAmelCase : Optional[int] = nn.Linear(3 , 4) __UpperCAmelCase : str = nn.BatchNormad(4) __UpperCAmelCase : int = nn.Linear(4 , 5) def a_ ( self : str , UpperCamelCase_ : List[str]): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(UpperCamelCase_))) class a__ ( unittest.TestCase ): def a_ ( self : Dict): """simple docstring""" __UpperCAmelCase : Optional[Any] = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , model.state_dict()) __UpperCAmelCase : Union[str, Any] = os.path.join(UpperCamelCase_ , "index.json") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: __UpperCAmelCase : Optional[int] = os.path.join(UpperCamelCase_ , F"{key}.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) # TODO: add tests on the fact weights are properly loaded def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : int = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: __UpperCAmelCase : List[Any] = torch.randn(2 , 3 , dtype=UpperCamelCase_) with TemporaryDirectory() as tmp_dir: __UpperCAmelCase : Tuple = offload_weight(UpperCamelCase_ , "weight" , UpperCamelCase_ , {}) __UpperCAmelCase : Dict = os.path.join(UpperCamelCase_ , "weight.dat") self.assertTrue(os.path.isfile(UpperCamelCase_)) self.assertDictEqual(UpperCamelCase_ , {"weight": {"shape": [2, 3], "dtype": str(UpperCamelCase_).split(".")[1]}}) __UpperCAmelCase : Optional[Any] = load_offloaded_weight(UpperCamelCase_ , index["weight"]) self.assertTrue(torch.equal(UpperCamelCase_ , UpperCamelCase_)) def a_ ( self : List[str]): """simple docstring""" __UpperCAmelCase : List[Any] = ModelForTest() __UpperCAmelCase : Optional[int] = model.state_dict() __UpperCAmelCase : List[str] = {k: v for k, v in state_dict.items() if "linear2" not in k} __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "linear2" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : List[str] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) __UpperCAmelCase : Optional[int] = {k: v for k, v in state_dict.items() if "weight" in k} __UpperCAmelCase : Optional[Any] = {k: v for k, v in state_dict.items() if "weight" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) __UpperCAmelCase : Optional[Any] = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) with TemporaryDirectory() as tmp_dir: offload_state_dict(UpperCamelCase_ , UpperCamelCase_) # Duplicates are removed __UpperCAmelCase : str = OffloadedWeightsLoader(state_dict=UpperCamelCase_ , save_folder=UpperCamelCase_) # Every key is there with the right value self.assertEqual(sorted(UpperCamelCase_) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(UpperCamelCase_ , weight_map[key])) def a_ ( self : Union[str, Any]): """simple docstring""" __UpperCAmelCase : Any = {"a.1": 0, "a.10": 1, "a.2": 2} __UpperCAmelCase : Union[str, Any] = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1": 0, "a.2": 2}) __UpperCAmelCase : int = {"a.1.a": 0, "a.10.a": 1, "a.2.a": 2} __UpperCAmelCase : int = extract_submodules_state_dict(UpperCamelCase_ , ["a.1", "a.2"]) self.assertDictEqual(UpperCamelCase_ , {"a.1.a": 0, "a.2.a": 2})
77
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { """configuration_deberta""": ["""DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DebertaConfig""", """DebertaOnnxConfig"""], """tokenization_deberta""": ["""DebertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ["""DebertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """DebertaForMaskedLM""", """DebertaForQuestionAnswering""", """DebertaForSequenceClassification""", """DebertaForTokenClassification""", """DebertaModel""", """DebertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDebertaForMaskedLM""", """TFDebertaForQuestionAnswering""", """TFDebertaForSequenceClassification""", """TFDebertaForTokenClassification""", """TFDebertaModel""", """TFDebertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
"""simple docstring""" def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : Dict = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __UpperCAmelCase : Union[str, Any] = n - k # Calculate C(n,k) for i in range(UpperCamelCase ): result *= n - i result //= i + 1 return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , UpperCamelCase ) // (node_count + 1) def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" if n < 0: raise ValueError("factorial() not defined for negative values" ) __UpperCAmelCase : Optional[Any] = 1 for i in range(1 , n + 1 ): result *= i return result def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return catalan_number(UpperCamelCase ) * factorial(UpperCamelCase ) if __name__ == "__main__": A = int(input("""Enter the number of nodes: """).strip() or 0) if node_count <= 0: raise ValueError("""We need some nodes to work with.""") print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
77
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: A = None A = logging.get_logger(__name__) A = """▁""" A = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} A = { """vocab_file""": {"""google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"""}, """tokenizer_file""": { """google/pegasus-xsum""": """https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json""" }, } A = { """google/pegasus-xsum""": 512, } class a__ ( __magic_name__ ): lowercase_ = VOCAB_FILES_NAMES lowercase_ = PRETRAINED_VOCAB_FILES_MAP lowercase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase_ = PegasusTokenizer lowercase_ = ["input_ids", "attention_mask"] def __init__( self : str , UpperCamelCase_ : str=None , UpperCamelCase_ : Optional[int]=None , UpperCamelCase_ : int="<pad>" , UpperCamelCase_ : Optional[Any]="</s>" , UpperCamelCase_ : Any="<unk>" , UpperCamelCase_ : Tuple="<mask_2>" , UpperCamelCase_ : Any="<mask_1>" , UpperCamelCase_ : Tuple=None , UpperCamelCase_ : str=103 , **UpperCamelCase_ : Optional[Any] , ): """simple docstring""" __UpperCAmelCase : Optional[int] = offset if additional_special_tokens is not None: if not isinstance(UpperCamelCase_ , UpperCamelCase_): raise TypeError( F"additional_special_tokens should be of type {type(UpperCamelCase_)}, but is" F" {type(UpperCamelCase_)}") __UpperCAmelCase : Any = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"<unk_{i}>" for i in range(len(UpperCamelCase_) , self.offset - 1) ] if len(set(UpperCamelCase_)) != len(UpperCamelCase_): raise ValueError( "Please make sure that the provided additional_special_tokens do not contain an incorrectly" F" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.") __UpperCAmelCase : str = additional_special_tokens_extended else: __UpperCAmelCase : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"<unk_{i}>" for i in range(2 , self.offset)] super().__init__( UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , pad_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , mask_token_sent=UpperCamelCase_ , offset=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , **UpperCamelCase_ , ) __UpperCAmelCase : Optional[int] = vocab_file __UpperCAmelCase : List[str] = False if not self.vocab_file else True def a_ ( self : Union[str, Any] , UpperCamelCase_ : Optional[int]): """simple docstring""" __UpperCAmelCase : int = set(self.all_special_ids) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens) + 3)): raise ValueError( "There should be 3 special tokens: mask_token, pad_token, and eos_token +" F" {len(self.additional_special_tokens)} additional_special_tokens, but got {all_special_ids}") return [1 if x in all_special_ids else 0 for x in seq] def a_ ( self : Union[str, Any] , UpperCamelCase_ : List , UpperCamelCase_ : Optional[List] = None , UpperCamelCase_ : bool = False): """simple docstring""" if already_has_special_tokens: return self._special_token_mask(UpperCamelCase_) elif token_ids_a is None: return self._special_token_mask(UpperCamelCase_) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a) + [1] def a_ ( self : int , UpperCamelCase_ : Union[str, Any] , UpperCamelCase_ : List[Any]=None): """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a_ ( self : Union[str, Any] , UpperCamelCase_ : str , UpperCamelCase_ : Optional[str] = None): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer.") if not os.path.isdir(UpperCamelCase_): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __UpperCAmelCase : List[str] = os.path.join( UpperCamelCase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(UpperCamelCase_): copyfile(self.vocab_file , UpperCamelCase_) return (out_vocab_file,)
77
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) A = { """configuration_trocr""": ["""TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TrOCRConfig"""], """processing_trocr""": ["""TrOCRProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ """TROCR_PRETRAINED_MODEL_ARCHIVE_LIST""", """TrOCRForCausalLM""", """TrOCRPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
77
1
"""simple docstring""" import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging A = logging.get_logger(__name__) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Any: """simple docstring""" __UpperCAmelCase : List[Any] = set() __UpperCAmelCase : List[Any] = [] def parse_line(UpperCamelCase ): for line in fp: if isinstance(UpperCamelCase , UpperCamelCase ): __UpperCAmelCase : int = line.decode("UTF-8" ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(" " ): # process a single warning and move it to `selected_warnings`. if len(UpperCamelCase ) > 0: __UpperCAmelCase : Optional[Any] = "\n".join(UpperCamelCase ) # Only keep the warnings specified in `targets` if any(f": {x}: " in warning for x in targets ): selected_warnings.add(UpperCamelCase ) buffer.clear() continue else: __UpperCAmelCase : Optional[int] = line.strip() buffer.append(UpperCamelCase ) if from_gh: for filename in os.listdir(UpperCamelCase ): __UpperCAmelCase : List[Any] = os.path.join(UpperCamelCase , UpperCamelCase ) if not os.path.isdir(UpperCamelCase ): # read the file if filename != "warnings.txt": continue with open(UpperCamelCase ) as fp: parse_line(UpperCamelCase ) else: try: with zipfile.ZipFile(UpperCamelCase ) as z: for filename in z.namelist(): if not os.path.isdir(UpperCamelCase ): # read the file if filename != "warnings.txt": continue with z.open(UpperCamelCase ) as fp: parse_line(UpperCamelCase ) except Exception: logger.warning( f"{artifact_path} is either an invalid zip file or something else wrong. This file is skipped." ) return selected_warnings def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase : Optional[Any] = set() __UpperCAmelCase : List[str] = [os.path.join(UpperCamelCase , UpperCamelCase ) for p in os.listdir(UpperCamelCase ) if (p.endswith(".zip" ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(UpperCamelCase , UpperCamelCase ) ) return selected_warnings if __name__ == "__main__": def _UpperCamelCase ( UpperCamelCase ) -> int: """simple docstring""" return values.split("," ) A = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") parser.add_argument( """--output_dir""", type=str, required=True, help="""Where to store the downloaded artifacts and other result files.""", ) parser.add_argument("""--token""", default=None, type=str, help="""A token that has actions:read permission.""") # optional parameters parser.add_argument( """--targets""", default="""DeprecationWarning,UserWarning,FutureWarning""", type=list_str, help="""Comma-separated list of target warning(s) which we want to extract.""", ) parser.add_argument( """--from_gh""", action="""store_true""", help="""If running from a GitHub action workflow and collecting warnings from its artifacts.""", ) A = parser.parse_args() A = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links A = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, """artifacts.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print("""=""" * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts A = extract_warnings(args.output_dir, args.targets) A = sorted(selected_warnings) with open(os.path.join(args.output_dir, """selected_warnings.json"""), """w""", encoding="""UTF-8""") as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
77
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a__ ( unittest.TestCase ): def __init__( self : Optional[Any] , UpperCamelCase_ : int , UpperCamelCase_ : Any=13 , UpperCamelCase_ : Optional[int]=3 , UpperCamelCase_ : int=224 , UpperCamelCase_ : int=30 , UpperCamelCase_ : str=400 , UpperCamelCase_ : List[Any]=True , UpperCamelCase_ : Optional[Any]=None , UpperCamelCase_ : Optional[int]=True , UpperCamelCase_ : Optional[int]=[0.5, 0.5, 0.5] , UpperCamelCase_ : Optional[Any]=[0.5, 0.5, 0.5] , ): """simple docstring""" __UpperCAmelCase : Tuple = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : List[Any] = parent __UpperCAmelCase : Tuple = batch_size __UpperCAmelCase : Tuple = num_channels __UpperCAmelCase : List[Any] = image_size __UpperCAmelCase : str = min_resolution __UpperCAmelCase : Tuple = max_resolution __UpperCAmelCase : Optional[Any] = do_resize __UpperCAmelCase : Any = size __UpperCAmelCase : Any = do_normalize __UpperCAmelCase : Any = image_mean __UpperCAmelCase : Optional[Any] = image_std def a_ ( self : str): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class a__ ( __magic_name__ , unittest.TestCase ): lowercase_ = ViTImageProcessor if is_vision_available() else None def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Optional[Any] = EfficientFormerImageProcessorTester(self) @property def a_ ( self : Union[str, Any]): """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Tuple = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(UpperCamelCase_ , "image_mean")) self.assertTrue(hasattr(UpperCamelCase_ , "image_std")) self.assertTrue(hasattr(UpperCamelCase_ , "do_normalize")) self.assertTrue(hasattr(UpperCamelCase_ , "do_resize")) self.assertTrue(hasattr(UpperCamelCase_ , "size")) def a_ ( self : Dict): """simple docstring""" pass def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCAmelCase : str = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , Image.Image) # Test not batched input __UpperCAmelCase : Optional[int] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Tuple): """simple docstring""" __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCAmelCase : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , numpify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , np.ndarray) # Test not batched input __UpperCAmelCase : Tuple = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Any = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def a_ ( self : Any): """simple docstring""" __UpperCAmelCase : Dict = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCAmelCase : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=UpperCamelCase_ , torchify=UpperCamelCase_) for image in image_inputs: self.assertIsInstance(UpperCamelCase_ , torch.Tensor) # Test not batched input __UpperCAmelCase : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[int] = image_processor(UpperCamelCase_ , return_tensors="pt").pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
77
1
"""simple docstring""" from __future__ import annotations from random import choice def _UpperCamelCase ( UpperCamelCase ) -> Optional[Any]: """simple docstring""" return choice(UpperCamelCase ) def _UpperCamelCase ( UpperCamelCase , UpperCamelCase ) -> int: """simple docstring""" __UpperCAmelCase : List[Any] = random_pivot(UpperCamelCase ) # partition based on pivot # linear time __UpperCAmelCase : Optional[Any] = [e for e in lst if e < pivot] __UpperCAmelCase : List[str] = [e for e in lst if e > pivot] # if we get lucky, pivot might be the element we want. # we can easily see this: # small (elements smaller than k) # + pivot (kth element) # + big (elements larger than k) if len(UpperCamelCase ) == k - 1: return pivot # pivot is in elements bigger than k elif len(UpperCamelCase ) < k - 1: return kth_number(UpperCamelCase , k - len(UpperCamelCase ) - 1 ) # pivot is in elements smaller than k else: return kth_number(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" from collections import namedtuple A = namedtuple("""from_to""", """from_ to""") A = { """cubicmeter""": from_to(1, 1), """litre""": from_to(0.001, 1_000), """kilolitre""": from_to(1, 1), """gallon""": from_to(0.00454, 264.172), """cubicyard""": from_to(0.76455, 1.30795), """cubicfoot""": from_to(0.028, 35.3147), """cup""": from_to(0.000236588, 4226.75), } def _UpperCamelCase ( UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> float: """simple docstring""" if from_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'from_type' value: {from_type!r} Supported values are:\n" + ", ".join(UpperCamelCase ) ) if to_type not in METRIC_CONVERSION: raise ValueError( f"Invalid 'to_type' value: {to_type!r}. Supported values are:\n" + ", ".join(UpperCamelCase ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
77
1