code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _UpperCAmelCase ( __snake_case, __snake_case, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =IFInpaintingSuperResolutionPipeline lowerCamelCase__ =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'width', 'height'} lowerCamelCase__ =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'original_image'} ) lowerCamelCase__ =PipelineTesterMixin.required_optional_params - {'latents'} def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return self._get_superresolution_dummy_components() def SCREAMING_SNAKE_CASE (self , a_ , a_=0 ): '''simple docstring''' if str(a_ ).startswith('''mps''' ): __snake_case : Tuple = torch.manual_seed(a_ ) else: __snake_case : str = torch.Generator(device=a_ ).manual_seed(a_ ) __snake_case : Union[str, Any] = floats_tensor((1, 3, 16, 16) , rng=random.Random(a_ ) ).to(a_ ) __snake_case : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(a_ ) ).to(a_ ) __snake_case : Union[str, Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(a_ ) ).to(a_ ) __snake_case : List[str] = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1E-1 ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self._test_save_load_local() def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
102
"""simple docstring""" import pytest import datasets # Import fixture modules as plugins SCREAMING_SNAKE_CASE : List[Any] = ["""tests.fixtures.files""", """tests.fixtures.hub""", """tests.fixtures.fsspec"""] def lowercase ( _snake_case : Optional[int] , _snake_case : Optional[int] ) ->Tuple: """simple docstring""" for item in items: if any(marker in item.keywords for marker in ['''integration''', '''unit'''] ): continue item.add_marker(pytest.mark.unit ) def lowercase ( _snake_case : List[str] ) ->Optional[int]: """simple docstring""" config.addinivalue_line('''markers''' , '''torchaudio_latest: mark test to run with torchaudio>=0.12''' ) @pytest.fixture(autouse=_snake_case ) def lowercase ( _snake_case : Optional[Any] , _snake_case : Dict ) ->Any: """simple docstring""" __snake_case : List[Any] = tmp_path_factory.getbasetemp() / '''cache''' __snake_case : int = test_hf_cache_home / '''datasets''' __snake_case : Tuple = test_hf_cache_home / '''metrics''' __snake_case : List[str] = test_hf_cache_home / '''modules''' monkeypatch.setattr('''datasets.config.HF_DATASETS_CACHE''' , str(_snake_case ) ) monkeypatch.setattr('''datasets.config.HF_METRICS_CACHE''' , str(_snake_case ) ) monkeypatch.setattr('''datasets.config.HF_MODULES_CACHE''' , str(_snake_case ) ) __snake_case : Optional[int] = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr('''datasets.config.DOWNLOADED_DATASETS_PATH''' , str(_snake_case ) ) __snake_case : Tuple = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(_snake_case ) ) @pytest.fixture(autouse=_snake_case , scope='''session''' ) def lowercase ( ) ->Any: """simple docstring""" datasets.disable_progress_bar() @pytest.fixture(autouse=_snake_case ) def lowercase ( _snake_case : Tuple ) ->Union[str, Any]: """simple docstring""" monkeypatch.setattr('''datasets.config.HF_UPDATE_DOWNLOAD_COUNTS''' , _snake_case ) @pytest.fixture def lowercase ( _snake_case : Any ) ->Optional[Any]: """simple docstring""" monkeypatch.setattr('''sqlalchemy.util.deprecations.SILENCE_UBER_WARNING''' , _snake_case )
102
1
"""simple docstring""" import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors SCREAMING_SNAKE_CASE = logging.getLogger(__name__) class UpperCAmelCase_ ( A_ ): lowercase__ = '''sequence-classification''' def __init__( self : str , snake_case_ : Any ) -> Tuple: '''simple docstring''' if type(snake_case_ ) == dict: A__ = Namespace(**snake_case_ ) A__ = glue_output_modes[hparams.task] A__ = glue_tasks_num_labels[hparams.task] super().__init__(snake_case_ , snake_case_ , self.mode ) def __magic_name__ ( self : List[str] , **snake_case_ : List[Any] ) -> Any: '''simple docstring''' return self.model(**snake_case_ ) def __magic_name__ ( self : List[str] , snake_case_ : Optional[int] , snake_case_ : Optional[Any] ) -> Optional[int]: '''simple docstring''' A__ = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: A__ = batch[2] if self.config.model_type in ["bert", "xlnet", "albert"] else None A__ = self(**snake_case_ ) A__ = outputs[0] A__ = self.trainer.lr_schedulers[0]["scheduler"] A__ = {"loss": loss, "rate": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def __magic_name__ ( self : Any ) -> Dict: '''simple docstring''' A__ = self.hparams A__ = processors[args.task]() A__ = processor.get_labels() for mode in ["train", "dev"]: A__ = self._feature_file(snake_case_ ) if os.path.exists(snake_case_ ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , snake_case_ ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) A__ = ( processor.get_dev_examples(args.data_dir ) if mode == "dev" else processor.get_train_examples(args.data_dir ) ) A__ = convert_examples_to_features( snake_case_ , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("Saving features into cached file %s" , snake_case_ ) torch.save(snake_case_ , snake_case_ ) def __magic_name__ ( self : Dict , snake_case_ : str , snake_case_ : int , snake_case_ : bool = False ) -> DataLoader: '''simple docstring''' A__ = "dev" if mode == "test" else mode A__ = self._feature_file(snake_case_ ) logger.info("Loading features from cached file %s" , snake_case_ ) A__ = torch.load(snake_case_ ) A__ = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) A__ = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) A__ = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) if self.hparams.glue_output_mode == "classification": A__ = torch.tensor([f.label for f in features] , dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": A__ = torch.tensor([f.label for f in features] , dtype=torch.float ) return DataLoader( TensorDataset(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) , batch_size=snake_case_ , shuffle=snake_case_ , ) def __magic_name__ ( self : int , snake_case_ : int , snake_case_ : int ) -> Optional[int]: '''simple docstring''' A__ = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: A__ = batch[2] if self.config.model_type in ["bert", "xlnet", "albert"] else None A__ = self(**snake_case_ ) A__, A__ = outputs[:2] A__ = logits.detach().cpu().numpy() A__ = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def __magic_name__ ( self : Union[str, Any] , snake_case_ : List[Any] ) -> tuple: '''simple docstring''' A__ = torch.stack([x["val_loss"] for x in outputs] ).mean().detach().cpu().item() A__ = np.concatenate([x["pred"] for x in outputs] , axis=0 ) if self.hparams.glue_output_mode == "classification": A__ = np.argmax(snake_case_ , axis=1 ) elif self.hparams.glue_output_mode == "regression": A__ = np.squeeze(snake_case_ ) A__ = np.concatenate([x["target"] for x in outputs] , axis=0 ) A__ = [[] for _ in range(out_label_ids.shape[0] )] A__ = [[] for _ in range(out_label_ids.shape[0] )] A__ = {**{"val_loss": val_loss_mean}, **compute_metrics(self.hparams.task , snake_case_ , snake_case_ )} A__ = dict(results.items() ) A__ = results return ret, preds_list, out_label_list def __magic_name__ ( self : Dict , snake_case_ : list ) -> dict: '''simple docstring''' A__, A__, A__ = self._eval_end(snake_case_ ) A__ = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def __magic_name__ ( self : str , snake_case_ : str ) -> dict: '''simple docstring''' A__, A__, A__ = self._eval_end(snake_case_ ) A__ = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def __magic_name__ ( snake_case_ : Union[str, Any] , snake_case_ : Dict ) -> Optional[Any]: '''simple docstring''' BaseTransformer.add_model_specific_args(snake_case_ , snake_case_ ) parser.add_argument( "--max_seq_length" , default=128 , type=snake_case_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--task" , default="" , type=snake_case_ , required=snake_case_ , help="The GLUE task to run" , ) parser.add_argument( "--gpus" , default=0 , type=snake_case_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser def _SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: A__ = argparse.ArgumentParser() add_generic_args(lowercase_ , os.getcwd() ) A__ = GLUETransformer.add_model_specific_args(lowercase_ , os.getcwd() ) A__ = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: A__ = os.path.join( "./results" , f"""{args.task}_{time.strftime("%Y%m%d_%H%M%S" )}""" , ) os.makedirs(args.output_dir ) A__ = GLUETransformer(lowercase_ ) A__ = generic_train(lowercase_ , lowercase_ ) # Optionally, predict on dev set and write to output_dir if args.do_predict: A__ = sorted(glob.glob(os.path.join(args.output_dir , "checkpoint-epoch=*.ckpt" ) , recursive=lowercase_ ) ) A__ = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(lowercase_ ) if __name__ == "__main__": main()
230
"""simple docstring""" SCREAMING_SNAKE_CASE = [ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] SCREAMING_SNAKE_CASE = [ 999, 976, 952, 928, 905, 882, 858, 857, 810, 762, 715, 714, 572, 429, 428, 286, 285, 238, 190, 143, 142, 118, 95, 71, 47, 24, 0, ] SCREAMING_SNAKE_CASE = [ 999, 988, 977, 966, 955, 944, 933, 922, 911, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 350, 300, 299, 266, 233, 200, 199, 179, 159, 140, 120, 100, 99, 88, 77, 66, 55, 44, 33, 22, 11, 0, ] SCREAMING_SNAKE_CASE = [ 999, 995, 992, 989, 985, 981, 978, 975, 971, 967, 964, 961, 957, 956, 951, 947, 942, 937, 933, 928, 923, 919, 914, 913, 908, 903, 897, 892, 887, 881, 876, 871, 870, 864, 858, 852, 846, 840, 834, 828, 827, 820, 813, 806, 799, 792, 785, 784, 777, 770, 763, 756, 749, 742, 741, 733, 724, 716, 707, 699, 698, 688, 677, 666, 656, 655, 645, 634, 623, 613, 612, 598, 584, 570, 569, 555, 541, 527, 526, 505, 484, 483, 462, 440, 439, 396, 395, 352, 351, 308, 307, 264, 263, 220, 219, 176, 132, 88, 44, 0, ] SCREAMING_SNAKE_CASE = [ 999, 997, 995, 992, 990, 988, 986, 984, 981, 979, 977, 975, 972, 970, 968, 966, 964, 961, 959, 957, 956, 954, 951, 949, 946, 944, 941, 939, 936, 934, 931, 929, 926, 924, 921, 919, 916, 914, 913, 910, 907, 905, 902, 899, 896, 893, 891, 888, 885, 882, 879, 877, 874, 871, 870, 867, 864, 861, 858, 855, 852, 849, 846, 843, 840, 837, 834, 831, 828, 827, 824, 821, 817, 814, 811, 808, 804, 801, 798, 795, 791, 788, 785, 784, 780, 777, 774, 770, 766, 763, 760, 756, 752, 749, 746, 742, 741, 737, 733, 730, 726, 722, 718, 714, 710, 707, 703, 699, 698, 694, 690, 685, 681, 677, 673, 669, 664, 660, 656, 655, 650, 646, 641, 636, 632, 627, 622, 618, 613, 612, 607, 602, 596, 591, 586, 580, 575, 570, 569, 563, 557, 551, 545, 539, 533, 527, 526, 519, 512, 505, 498, 491, 484, 483, 474, 466, 457, 449, 440, 439, 428, 418, 407, 396, 395, 381, 366, 352, 351, 330, 308, 307, 286, 264, 263, 242, 220, 219, 176, 175, 132, 131, 88, 44, 0, ] SCREAMING_SNAKE_CASE = [ 999, 991, 982, 974, 966, 958, 950, 941, 933, 925, 916, 908, 900, 899, 874, 850, 825, 800, 799, 700, 600, 500, 400, 300, 200, 100, 0, ] SCREAMING_SNAKE_CASE = [ 999, 992, 985, 978, 971, 964, 957, 949, 942, 935, 928, 921, 914, 907, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 300, 299, 200, 199, 100, 99, 0, ] SCREAMING_SNAKE_CASE = [ 999, 996, 992, 989, 985, 982, 979, 975, 972, 968, 965, 961, 958, 955, 951, 948, 944, 941, 938, 934, 931, 927, 924, 920, 917, 914, 910, 907, 903, 900, 899, 891, 884, 876, 869, 861, 853, 846, 838, 830, 823, 815, 808, 800, 799, 788, 777, 766, 755, 744, 733, 722, 711, 700, 699, 688, 677, 666, 655, 644, 633, 622, 611, 600, 599, 585, 571, 557, 542, 528, 514, 500, 499, 485, 471, 457, 442, 428, 414, 400, 399, 379, 359, 340, 320, 300, 299, 279, 259, 240, 220, 200, 199, 166, 133, 100, 99, 66, 33, 0, ]
230
1
class _SCREAMING_SNAKE_CASE : def __init__( self , lowercase , lowercase , lowercase ) -> Union[str, Any]: lowerCamelCase_ = None lowerCamelCase_ = None lowerCamelCase_ = graph self._normalize_graph(lowercase , lowercase ) lowerCamelCase_ = len(lowercase ) lowerCamelCase_ = None def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> List[str]: if sources is int: lowerCamelCase_ = [sources] if sinks is int: lowerCamelCase_ = [sinks] if len(lowercase ) == 0 or len(lowercase ) == 0: return lowerCamelCase_ = sources[0] lowerCamelCase_ = sinks[0] # make fake vertex if there are more # than one source or sink if len(lowercase ) > 1 or len(lowercase ) > 1: lowerCamelCase_ = 0 for i in sources: max_input_flow += sum(self.graph[i] ) lowerCamelCase_ = len(self.graph ) + 1 for room in self.graph: room.insert(0 , 0 ) self.graph.insert(0 , [0] * size ) for i in sources: lowerCamelCase_ = max_input_flow lowerCamelCase_ = 0 lowerCamelCase_ = len(self.graph ) + 1 for room in self.graph: room.append(0 ) self.graph.append([0] * size ) for i in sinks: lowerCamelCase_ = max_input_flow lowerCamelCase_ = size - 1 def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: if self.maximum_flow_algorithm is None: raise Exception("You need to set maximum flow algorithm before." ) if self.source_index is None or self.sink_index is None: return 0 self.maximum_flow_algorithm.execute() return self.maximum_flow_algorithm.getMaximumFlow() def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Optional[Any]: lowerCamelCase_ = algorithm(self ) class _SCREAMING_SNAKE_CASE : def __init__( self , lowercase ) -> List[Any]: lowerCamelCase_ = flow_network lowerCamelCase_ = flow_network.verticesCount lowerCamelCase_ = flow_network.sourceIndex lowerCamelCase_ = flow_network.sinkIndex # it's just a reference, so you shouldn't change # it in your algorithms, use deep copy before doing that lowerCamelCase_ = flow_network.graph lowerCamelCase_ = False def SCREAMING_SNAKE_CASE_( self ) -> Tuple: if not self.executed: self._algorithm() lowerCamelCase_ = True def SCREAMING_SNAKE_CASE_( self ) -> List[str]: pass class _SCREAMING_SNAKE_CASE ( snake_case_ ): def __init__( self , lowercase ) -> Any: super().__init__(lowercase ) # use this to save your result lowerCamelCase_ = -1 def SCREAMING_SNAKE_CASE_( self ) -> Any: if not self.executed: raise Exception("You should execute algorithm before using its result!" ) return self.maximum_flow class _SCREAMING_SNAKE_CASE ( snake_case_ ): def __init__( self , lowercase ) -> str: super().__init__(lowercase ) lowerCamelCase_ = [[0] * self.verticies_count for i in range(self.verticies_count )] lowerCamelCase_ = [0] * self.verticies_count lowerCamelCase_ = [0] * self.verticies_count def SCREAMING_SNAKE_CASE_( self ) -> str: lowerCamelCase_ = self.verticies_count # push some substance to graph for nextvertex_index, bandwidth in enumerate(self.graph[self.source_index] ): self.preflow[self.source_index][nextvertex_index] += bandwidth self.preflow[nextvertex_index][self.source_index] -= bandwidth self.excesses[nextvertex_index] += bandwidth # Relabel-to-front selection rule lowerCamelCase_ = [ i for i in range(self.verticies_count ) if i != self.source_index and i != self.sink_index ] # move through list lowerCamelCase_ = 0 while i < len(lowercase ): lowerCamelCase_ = vertices_list[i] lowerCamelCase_ = self.heights[vertex_index] self.process_vertex(lowercase ) if self.heights[vertex_index] > previous_height: # if it was relabeled, swap elements # and start from 0 index vertices_list.insert(0 , vertices_list.pop(lowercase ) ) lowerCamelCase_ = 0 else: i += 1 lowerCamelCase_ = sum(self.preflow[self.source_index] ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> List[Any]: while self.excesses[vertex_index] > 0: for neighbour_index in range(self.verticies_count ): # if it's neighbour and current vertex is higher if ( self.graph[vertex_index][neighbour_index] - self.preflow[vertex_index][neighbour_index] > 0 and self.heights[vertex_index] > self.heights[neighbour_index] ): self.push(lowercase , lowercase ) self.relabel(lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> Any: lowerCamelCase_ = min( self.excesses[from_index] , self.graph[from_index][to_index] - self.preflow[from_index][to_index] , ) self.preflow[from_index][to_index] += preflow_delta self.preflow[to_index][from_index] -= preflow_delta self.excesses[from_index] -= preflow_delta self.excesses[to_index] += preflow_delta def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Dict: lowerCamelCase_ = None for to_index in range(self.verticies_count ): if ( self.graph[vertex_index][to_index] - self.preflow[vertex_index][to_index] > 0 ) and (min_height is None or self.heights[to_index] < min_height): lowerCamelCase_ = self.heights[to_index] if min_height is not None: lowerCamelCase_ = min_height + 1 if __name__ == "__main__": __A =[0] __A =[3] # graph = [ # [0, 0, 4, 6, 0, 0], # [0, 0, 5, 2, 0, 0], # [0, 0, 0, 0, 4, 4], # [0, 0, 0, 0, 6, 6], # [0, 0, 0, 0, 0, 0], # [0, 0, 0, 0, 0, 0], # ] __A =[[0, 7, 0, 0], [0, 0, 6, 0], [0, 0, 0, 8], [9, 0, 0, 0]] # prepare our network __A =FlowNetwork(graph, entrances, exits) # set algorithm flow_network.set_maximum_flow_algorithm(PushRelabelExecutor) # and calculate __A =flow_network.find_maximum_flow() print(F"""maximum flow is {maximum_flow}""")
19
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def a ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=10_24 , lowerCamelCase__=10_24 , lowerCamelCase__=False , **lowerCamelCase__ ): '''simple docstring''' A_ : Any = AutoTokenizer.from_pretrained(lowerCamelCase__ ) A_ : Union[str, Any] = SeqaSeqDataset(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , type_path="""train""" , **lowerCamelCase__ ) A_ : Optional[Any] = tok.pad_token_id def get_lens(lowerCamelCase__ ): A_ : int = tqdm( DataLoader(lowerCamelCase__ , batch_size=5_12 , num_workers=8 , shuffle=lowerCamelCase__ , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) A_ : int = [] for batch in dl: A_ : str = batch["""input_ids"""].ne(lowerCamelCase__ ).sum(1 ).tolist() A_ : Tuple = batch["""labels"""].ne(lowerCamelCase__ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(lowerCamelCase__ , lowerCamelCase__ ): max_lens.append(max(lowerCamelCase__ , lowerCamelCase__ ) ) else: max_lens.extend(lowerCamelCase__ ) return max_lens A_ : str = get_lens(lowerCamelCase__ ) A_ : Dict = SeqaSeqDataset(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , type_path="""val""" , **lowerCamelCase__ ) A_ : List[Any] = get_lens(lowerCamelCase__ ) pickle_save(lowerCamelCase__ , train_ds.len_file ) pickle_save(lowerCamelCase__ , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
206
0
'''simple docstring''' import timeit import numpy as np import datasets from datasets.arrow_writer import ArrowWriter from datasets.features.features import _ArrayXD def _snake_case ( A ) -> List[Any]: def wrapper(*A , **A ): lowerCAmelCase__ = timeit.default_timer() lowerCAmelCase__ = func(*A , **A ) lowerCAmelCase__ = timeit.default_timer() - starttime return delta lowerCAmelCase__ = func.__name__ return wrapper def _snake_case ( A , A=100 , A=None ) -> Dict: lowerCAmelCase__ = [] lowerCAmelCase__ = seq_shapes or {} for i in range(A ): lowerCAmelCase__ = {} for col_id, (k, v) in enumerate(features.items() ): if isinstance(A , _ArrayXD ): lowerCAmelCase__ = np.random.rand(*v.shape ).astype(v.dtype ) elif isinstance(A , datasets.Value ): if v.dtype == "string": lowerCAmelCase__ = '''The small grey turtle was surprisingly fast when challenged.''' else: lowerCAmelCase__ = np.random.randint(10 , size=1 ).astype(v.dtype ).item() elif isinstance(A , datasets.Sequence ): while isinstance(A , datasets.Sequence ): lowerCAmelCase__ = v.feature lowerCAmelCase__ = seq_shapes[k] lowerCAmelCase__ = np.random.rand(*A ).astype(v.dtype ) lowerCAmelCase__ = data dummy_data.append((i, example) ) return dummy_data def _snake_case ( A , A , A=100 , A=None ) -> Tuple: lowerCAmelCase__ = generate_examples(A , num_examples=A , seq_shapes=A ) with ArrowWriter(features=A , path=A ) as writer: for key, record in dummy_data: lowerCAmelCase__ = features.encode_example(A ) writer.write(A ) lowerCAmelCase__ , lowerCAmelCase__ = writer.finalize() if not num_final_examples == num_examples: raise ValueError( F"""Error writing the dataset, wrote {num_final_examples} examples but should have written {num_examples}.""" ) lowerCAmelCase__ = datasets.Dataset.from_file(filename=A , info=datasets.DatasetInfo(features=A ) ) return dataset
228
'''simple docstring''' from __future__ import annotations def _snake_case ( A , A ) -> float: lowerCAmelCase__ = sorted(numsa + numsa ) lowerCAmelCase__ , lowerCAmelCase__ = divmod(len(A ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase = [float(x) for x in input('''Enter the elements of first array: ''').split()] __UpperCAmelCase = [float(x) for x in input('''Enter the elements of second array: ''').split()] print(f"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
228
1
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class UpperCAmelCase_ ( UpperCamelCase , unittest.TestCase ): '''simple docstring''' __A : str = CodeGenTokenizer __A : Dict = CodeGenTokenizerFast __A : List[Any] = True __A : List[str] = {"add_prefix_space": True} __A : List[str] = False def _snake_case ( self ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCamelCase : str = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", "<|endoftext|>", ] lowerCamelCase : Dict = dict(zip(__A , range(len(__A ) ) ) ) lowerCamelCase : List[str] = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCamelCase : List[str] = {"unk_token": "<unk>"} lowerCamelCase : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) lowerCamelCase : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(__A ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(__A ) ) def _snake_case ( self , **__A ): """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **__A ) def _snake_case ( self , **__A ): """simple docstring""" kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **__A ) def _snake_case ( self , __A ): """simple docstring""" lowerCamelCase : List[str] = "lower newer" lowerCamelCase : Optional[int] = "lower newer" return input_text, output_text def _snake_case ( self ): """simple docstring""" lowerCamelCase : str = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) lowerCamelCase : Dict = "lower newer" lowerCamelCase : Any = ["\u0120low", "er", "\u0120", "n", "e", "w", "er"] lowerCamelCase : int = tokenizer.tokenize(__A , add_prefix_space=__A ) self.assertListEqual(__A , __A ) lowerCamelCase : Tuple = tokens + [tokenizer.unk_token] lowerCamelCase : Tuple = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) def _snake_case ( self ): """simple docstring""" if not self.test_rust_tokenizer: return lowerCamelCase : int = self.get_tokenizer() lowerCamelCase : List[Any] = self.get_rust_tokenizer(add_prefix_space=__A ) lowerCamelCase : int = "lower newer" # Testing tokenization lowerCamelCase : Union[str, Any] = tokenizer.tokenize(__A , add_prefix_space=__A ) lowerCamelCase : Optional[int] = rust_tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) # Testing conversion to ids without special tokens lowerCamelCase : Optional[Any] = tokenizer.encode(__A , add_special_tokens=__A , add_prefix_space=__A ) lowerCamelCase : List[Any] = rust_tokenizer.encode(__A , add_special_tokens=__A ) self.assertListEqual(__A , __A ) # Testing conversion to ids with special tokens lowerCamelCase : int = self.get_rust_tokenizer(add_prefix_space=__A ) lowerCamelCase : Dict = tokenizer.encode(__A , add_prefix_space=__A ) lowerCamelCase : Dict = rust_tokenizer.encode(__A ) self.assertListEqual(__A , __A ) # Testing the unknown token lowerCamelCase : Any = tokens + [rust_tokenizer.unk_token] lowerCamelCase : str = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(__A ) , __A ) def _snake_case ( self , *__A , **__A ): """simple docstring""" pass def _snake_case ( self , __A=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): lowerCamelCase : str = self.rust_tokenizer_class.from_pretrained(__A , **__A ) # Simple input lowerCamelCase : Dict = "This is a simple input" lowerCamelCase : List[Any] = ["This is a simple input 1", "This is a simple input 2"] lowerCamelCase : int = ("This is a simple input", "This is a pair") lowerCamelCase : Optional[int] = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="max_length" ) # Simple input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="max_length" ) # Simple input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="max_length" , ) # Pair input self.assertRaises(__A , tokenizer_r.encode , __A , max_length=__A , padding="max_length" ) # Pair input self.assertRaises(__A , tokenizer_r.encode_plus , __A , max_length=__A , padding="max_length" ) # Pair input self.assertRaises( __A , tokenizer_r.batch_encode_plus , __A , max_length=__A , padding="max_length" , ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[str] = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token="<pad>" ) # Simple input lowerCamelCase : Optional[int] = "This is a simple input" lowerCamelCase : int = ["This is a simple input looooooooong", "This is a simple input"] lowerCamelCase : List[Any] = ("This is a simple input", "This is a pair") lowerCamelCase : Union[str, Any] = [ ("This is a simple input loooooong", "This is a simple input"), ("This is a simple pair loooooong", "This is a simple pair"), ] lowerCamelCase : Any = tokenizer.pad_token_id lowerCamelCase : Any = tokenizer(__A , padding="max_length" , max_length=30 , return_tensors="np" ) lowerCamelCase : Any = tokenizer(__A , padding=__A , truncate=__A , return_tensors="np" ) lowerCamelCase : List[str] = tokenizer(*__A , padding="max_length" , max_length=60 , return_tensors="np" ) lowerCamelCase : Tuple = tokenizer(__A , padding=__A , truncate=__A , return_tensors="np" ) # s # test single string max_length padding self.assertEqual(out_s["input_ids"].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s["input_ids"] ) self.assertTrue(0 in out_s["attention_mask"] ) # s2 # test automatic padding self.assertEqual(out_sa["input_ids"].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["input_ids"][0] ) self.assertFalse(0 in out_sa["attention_mask"][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["input_ids"][1] ) self.assertTrue(0 in out_sa["attention_mask"][1] ) # p # test single pair max_length padding self.assertEqual(out_p["input_ids"].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p["input_ids"] ) self.assertTrue(0 in out_p["attention_mask"] ) # p2 # test automatic padding pair self.assertEqual(out_pa["input_ids"].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["input_ids"][0] ) self.assertFalse(0 in out_pa["attention_mask"][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["input_ids"][1] ) self.assertTrue(0 in out_pa["attention_mask"][1] ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = "$$$" lowerCamelCase : int = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=__A , add_bos_token=__A ) lowerCamelCase : Any = "This is a simple input" lowerCamelCase : int = ["This is a simple input 1", "This is a simple input 2"] lowerCamelCase : Dict = tokenizer.bos_token_id lowerCamelCase : Tuple = tokenizer(__A ) lowerCamelCase : List[str] = tokenizer(__A ) self.assertEqual(out_s.input_ids[0] , __A ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) lowerCamelCase : int = tokenizer.decode(out_s.input_ids ) lowerCamelCase : List[Any] = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , __A ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def _snake_case ( self ): """simple docstring""" lowerCamelCase : Optional[int] = CodeGenTokenizer.from_pretrained("Salesforce/codegen-350M-mono" ) lowerCamelCase : str = "\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#" lowerCamelCase : Union[str, Any] = "\nif len_a > len_b: result = a\nelse: result = b" lowerCamelCase : Any = tokenizer.encode(__A ) lowerCamelCase : Union[str, Any] = ["^#", re.escape("<|endoftext|>" ), "^'''", "^\"\"\"", "\n\n\n"] lowerCamelCase : int = tokenizer.decode(__A , truncate_before_pattern=__A ) self.assertEqual(__A , __A ) def _snake_case ( self ): """simple docstring""" pass
283
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class UpperCAmelCase_ ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' __A : Tuple = StableUnCLIPPipeline __A : Optional[int] = TEXT_TO_IMAGE_PARAMS __A : str = TEXT_TO_IMAGE_BATCH_PARAMS __A : int = TEXT_TO_IMAGE_IMAGE_PARAMS __A : Tuple = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false __A : Union[str, Any] = False def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[str] = 32 lowerCamelCase : Dict = embedder_hidden_size # prior components torch.manual_seed(0 ) lowerCamelCase : Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) lowerCamelCase : Optional[int] = CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__A , projection_dim=__A , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) lowerCamelCase : List[Any] = PriorTransformer( num_attention_heads=2 , attention_head_dim=12 , embedding_dim=__A , num_layers=1 , ) torch.manual_seed(0 ) lowerCamelCase : Dict = DDPMScheduler( variance_type="fixed_small_log" , prediction_type="sample" , num_train_timesteps=1000 , clip_sample=__A , clip_sample_range=5.0 , beta_schedule="squaredcos_cap_v2" , ) # regular denoising components torch.manual_seed(0 ) lowerCamelCase : Optional[int] = StableUnCLIPImageNormalizer(embedding_dim=__A ) lowerCamelCase : Tuple = DDPMScheduler(beta_schedule="squaredcos_cap_v2" ) torch.manual_seed(0 ) lowerCamelCase : List[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) torch.manual_seed(0 ) lowerCamelCase : str = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__A , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) ) torch.manual_seed(0 ) lowerCamelCase : Any = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock2D", "DownBlock2D") , up_block_types=("UpBlock2D", "CrossAttnUpBlock2D") , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type="projection" , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__A , layers_per_block=1 , upcast_attention=__A , use_linear_projection=__A , ) torch.manual_seed(0 ) lowerCamelCase : int = DDIMScheduler( beta_schedule="scaled_linear" , beta_start=0.00085 , beta_end=0.012 , prediction_type="v_prediction" , set_alpha_to_one=__A , steps_offset=1 , ) torch.manual_seed(0 ) lowerCamelCase : Optional[Any] = AutoencoderKL() lowerCamelCase : Optional[int] = { # prior components "prior_tokenizer": prior_tokenizer, "prior_text_encoder": prior_text_encoder, "prior": prior, "prior_scheduler": prior_scheduler, # image noising components "image_normalizer": image_normalizer, "image_noising_scheduler": image_noising_scheduler, # regular denoising components "tokenizer": tokenizer, "text_encoder": text_encoder, "unet": unet, "scheduler": scheduler, "vae": vae, } return components def _snake_case ( self , __A , __A=0 ): """simple docstring""" if str(__A ).startswith("mps" ): lowerCamelCase : Optional[int] = torch.manual_seed(__A ) else: lowerCamelCase : Optional[Any] = torch.Generator(device=__A ).manual_seed(__A ) lowerCamelCase : Tuple = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "prior_num_inference_steps": 2, "output_type": "numpy", } return inputs def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = torch_device == "cpu" self._test_attention_slicing_forward_pass(test_max_difference=__A ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[Any] = torch_device in ["cpu", "mps"] self._test_inference_batch_single_identical(test_max_difference=__A ) @slow @require_torch_gpu class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case ( self ): """simple docstring""" lowerCamelCase : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy" ) lowerCamelCase : str = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCamelCase : List[Any] = torch.Generator(device="cpu" ).manual_seed(0 ) lowerCamelCase : Dict = pipe("anime turle" , generator=__A , output_type="np" ) lowerCamelCase : Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__A , __A ) def _snake_case ( self ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowerCamelCase : int = StableUnCLIPPipeline.from_pretrained("fusing/stable-unclip-2-1-l" , torch_dtype=torch.floataa ) lowerCamelCase : Union[str, Any] = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() lowerCamelCase : Any = pipe( "anime turtle" , prior_num_inference_steps=2 , num_inference_steps=2 , output_type="np" , ) lowerCamelCase : List[str] = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
283
1
def lowerCamelCase_ ( _a , _a ): """simple docstring""" lowerCAmelCase__ : Tuple = len(__lowerCAmelCase ) lowerCAmelCase__ : Tuple = len(__lowerCAmelCase ) lowerCAmelCase__ : int = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] lowerCAmelCase__ : Union[str, Any] = True for i in range(__lowerCAmelCase ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: lowerCAmelCase__ : Optional[Any] = True if a[i].islower(): lowerCAmelCase__ : Optional[int] = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
362
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { '''facebook/xlm-roberta-xl''': '''https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json''', '''facebook/xlm-roberta-xxl''': '''https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json''', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class _a ( _lowercase): _a : Optional[int] = '''xlm-roberta-xl''' def __init__( self : Any , _SCREAMING_SNAKE_CASE : str=25_0880 , _SCREAMING_SNAKE_CASE : Optional[Any]=2560 , _SCREAMING_SNAKE_CASE : int=36 , _SCREAMING_SNAKE_CASE : Optional[int]=32 , _SCREAMING_SNAKE_CASE : Any=1_0240 , _SCREAMING_SNAKE_CASE : List[str]="gelu" , _SCREAMING_SNAKE_CASE : Optional[Any]=0.1 , _SCREAMING_SNAKE_CASE : Dict=0.1 , _SCREAMING_SNAKE_CASE : Optional[int]=514 , _SCREAMING_SNAKE_CASE : Optional[int]=1 , _SCREAMING_SNAKE_CASE : Tuple=0.02 , _SCREAMING_SNAKE_CASE : Dict=1E-05 , _SCREAMING_SNAKE_CASE : Tuple=1 , _SCREAMING_SNAKE_CASE : Optional[int]=0 , _SCREAMING_SNAKE_CASE : Tuple=2 , _SCREAMING_SNAKE_CASE : Union[str, Any]="absolute" , _SCREAMING_SNAKE_CASE : Any=True , _SCREAMING_SNAKE_CASE : Any=None , **_SCREAMING_SNAKE_CASE : Tuple , )-> str: super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : Any = vocab_size lowerCAmelCase__ : List[Any] = hidden_size lowerCAmelCase__ : int = num_hidden_layers lowerCAmelCase__ : Dict = num_attention_heads lowerCAmelCase__ : str = hidden_act lowerCAmelCase__ : List[Any] = intermediate_size lowerCAmelCase__ : Tuple = hidden_dropout_prob lowerCAmelCase__ : str = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = max_position_embeddings lowerCAmelCase__ : Any = type_vocab_size lowerCAmelCase__ : Optional[Any] = initializer_range lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : Optional[int] = position_embedding_type lowerCAmelCase__ : Any = use_cache lowerCAmelCase__ : List[Any] = classifier_dropout class _a ( _lowercase): @property def UpperCAmelCase__( self : Any )-> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCAmelCase__ : int = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCAmelCase__ : List[Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
211
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { '''microsoft/trocr-base-handwritten''': ( '''https://huggingface.co/microsoft/trocr-base-handwritten/resolve/main/config.json''' ), # See all TrOCR models at https://huggingface.co/models?filter=trocr } class _snake_case ( _lowercase ): lowerCamelCase__: List[Any] = "trocr" lowerCamelCase__: Any = ["past_key_values"] lowerCamelCase__: Any = { "num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model", "num_hidden_layers": "decoder_layers", } def __init__( self: Dict , __lowerCamelCase: Union[str, Any]=5_02_65 , __lowerCamelCase: Any=10_24 , __lowerCamelCase: Union[str, Any]=12 , __lowerCamelCase: Optional[int]=16 , __lowerCamelCase: Optional[Any]=40_96 , __lowerCamelCase: str="gelu" , __lowerCamelCase: Optional[int]=5_12 , __lowerCamelCase: Dict=0.1 , __lowerCamelCase: List[str]=0.0 , __lowerCamelCase: Tuple=0.0 , __lowerCamelCase: str=2 , __lowerCamelCase: Dict=0.02 , __lowerCamelCase: Any=0.0 , __lowerCamelCase: List[str]=True , __lowerCamelCase: int=False , __lowerCamelCase: List[str]=True , __lowerCamelCase: List[str]=True , __lowerCamelCase: Dict=1 , __lowerCamelCase: Union[str, Any]=0 , __lowerCamelCase: Dict=2 , **__lowerCamelCase: str , ) -> Optional[int]: __UpperCAmelCase : List[Any] = vocab_size __UpperCAmelCase : Tuple = d_model __UpperCAmelCase : Tuple = decoder_layers __UpperCAmelCase : List[Any] = decoder_attention_heads __UpperCAmelCase : List[Any] = decoder_ffn_dim __UpperCAmelCase : List[str] = activation_function __UpperCAmelCase : Dict = max_position_embeddings __UpperCAmelCase : str = dropout __UpperCAmelCase : str = attention_dropout __UpperCAmelCase : List[Any] = activation_dropout __UpperCAmelCase : Union[str, Any] = init_std __UpperCAmelCase : Dict = decoder_layerdrop __UpperCAmelCase : Any = use_cache __UpperCAmelCase : Optional[Any] = scale_embedding __UpperCAmelCase : Optional[int] = use_learned_position_embeddings __UpperCAmelCase : List[str] = layernorm_embedding super().__init__( pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , **__lowerCamelCase , )
157
def _UpperCamelCase ( snake_case__ ) -> list: __UpperCAmelCase : Dict = [0] * len(snake_case__ ) for i in range(1, len(snake_case__ ) ): # use last results for better performance - dynamic programming __UpperCAmelCase : Any = prefix_result[i - 1] while j > 0 and input_string[i] != input_string[j]: __UpperCAmelCase : Union[str, Any] = prefix_result[j - 1] if input_string[i] == input_string[j]: j += 1 __UpperCAmelCase : Tuple = j return prefix_result def _UpperCamelCase ( snake_case__ ) -> int: return max(prefix_function(snake_case__ ) ) if __name__ == "__main__": import doctest doctest.testmod()
157
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowercase : Optional[Any] = logging.get_logger(__name__) __lowercase : List[str] = { '''facebook/xmod-base''': '''https://huggingface.co/facebook/xmod-base/resolve/main/config.json''', '''facebook/xmod-large-prenorm''': '''https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json''', '''facebook/xmod-base-13-125k''': '''https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json''', '''facebook/xmod-base-30-125k''': '''https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json''', '''facebook/xmod-base-30-195k''': '''https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json''', '''facebook/xmod-base-60-125k''': '''https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json''', '''facebook/xmod-base-60-265k''': '''https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json''', '''facebook/xmod-base-75-125k''': '''https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json''', '''facebook/xmod-base-75-269k''': '''https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json''', } class _lowercase ( __a ): """simple docstring""" lowercase__ = '''xmod''' def __init__( self : int , UpperCamelCase__ : Optional[int]=30522 , UpperCamelCase__ : List[Any]=768 , UpperCamelCase__ : List[str]=12 , UpperCamelCase__ : Dict=12 , UpperCamelCase__ : Dict=3072 , UpperCamelCase__ : List[Any]="gelu" , UpperCamelCase__ : Tuple=0.1 , UpperCamelCase__ : Tuple=0.1 , UpperCamelCase__ : Optional[Any]=512 , UpperCamelCase__ : List[str]=2 , UpperCamelCase__ : Tuple=0.02 , UpperCamelCase__ : Tuple=1E-12 , UpperCamelCase__ : Tuple=1 , UpperCamelCase__ : Dict=0 , UpperCamelCase__ : int=2 , UpperCamelCase__ : Tuple="absolute" , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : Optional[int]=False , UpperCamelCase__ : Tuple=2 , UpperCamelCase__ : Optional[int]=False , UpperCamelCase__ : int=True , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : Optional[int]=("en_XX",) , UpperCamelCase__ : int=None , **UpperCamelCase__ : Union[str, Any] , ) -> Dict: '''simple docstring''' super().__init__(pad_token_id=UpperCamelCase__ , bos_token_id=UpperCamelCase__ , eos_token_id=UpperCamelCase__ , **UpperCamelCase__ ) __UpperCamelCase =vocab_size __UpperCamelCase =hidden_size __UpperCamelCase =num_hidden_layers __UpperCamelCase =num_attention_heads __UpperCamelCase =hidden_act __UpperCamelCase =intermediate_size __UpperCamelCase =hidden_dropout_prob __UpperCamelCase =attention_probs_dropout_prob __UpperCamelCase =max_position_embeddings __UpperCamelCase =type_vocab_size __UpperCamelCase =initializer_range __UpperCamelCase =layer_norm_eps __UpperCamelCase =position_embedding_type __UpperCamelCase =use_cache __UpperCamelCase =classifier_dropout __UpperCamelCase =pre_norm __UpperCamelCase =adapter_reduction_factor __UpperCamelCase =adapter_layer_norm __UpperCamelCase =adapter_reuse_layer_norm __UpperCamelCase =ln_before_adapter __UpperCamelCase =list(UpperCamelCase__ ) __UpperCamelCase =default_language class _lowercase ( __a ): """simple docstring""" @property def UpperCAmelCase_ ( self : Optional[int] ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": __UpperCamelCase ={0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: __UpperCamelCase ={0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
366
"""simple docstring""" import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( __a , unittest.TestCase ): """simple docstring""" lowercase__ = LongformerTokenizer lowercase__ = True lowercase__ = LongformerTokenizerFast lowercase__ = True def UpperCAmelCase_ ( self : Optional[Any] ) -> Any: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCamelCase =[ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __UpperCamelCase =dict(zip(UpperCamelCase__ , range(len(UpperCamelCase__ ) ) ) ) __UpperCamelCase =['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __UpperCamelCase ={'''unk_token''': '''<unk>'''} __UpperCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __UpperCamelCase =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(UpperCamelCase__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(UpperCamelCase__ ) ) def UpperCAmelCase_ ( self : Optional[int] , **UpperCamelCase__ : str ) -> Dict: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[str] , **UpperCamelCase__ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[str] , UpperCamelCase__ : List[str] ) -> Optional[Any]: '''simple docstring''' __UpperCamelCase ='''lower newer''' __UpperCamelCase ='''lower newer''' return input_text, output_text def UpperCAmelCase_ ( self : int ) -> List[Any]: '''simple docstring''' __UpperCamelCase =self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) __UpperCamelCase ='''lower newer''' __UpperCamelCase =['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] __UpperCamelCase =tokenizer.tokenize(UpperCamelCase__ ) # , add_prefix_space=True) self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =tokens + [tokenizer.unk_token] __UpperCamelCase =[0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) , UpperCamelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> int: '''simple docstring''' __UpperCamelCase =self.get_tokenizer() self.assertListEqual(tokenizer.encode('''Hello world!''' , add_special_tokens=UpperCamelCase__ ) , [0, 31414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode('''Hello world! cécé herlolip 418''' , add_special_tokens=UpperCamelCase__ ) , [0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2] , ) @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' __UpperCamelCase =self.tokenizer_class.from_pretrained('''allenai/longformer-base-4096''' ) __UpperCamelCase =tokenizer.encode('''sequence builders''' , add_special_tokens=UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode('''multi-sequence build''' , add_special_tokens=UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode( '''sequence builders''' , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.build_inputs_with_special_tokens(UpperCamelCase__ ) __UpperCamelCase =tokenizer.build_inputs_with_special_tokens(UpperCamelCase__ , UpperCamelCase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def UpperCAmelCase_ ( self : int ) -> Dict: '''simple docstring''' __UpperCamelCase =self.get_tokenizer() __UpperCamelCase ='''Encode this sequence.''' __UpperCamelCase =tokenizer.byte_encoder[''' '''.encode('''utf-8''' )[0]] # Testing encoder arguments __UpperCamelCase =tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) tokenizer.add_special_tokens({'''bos_token''': '''<s>'''} ) __UpperCamelCase =tokenizer.encode(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(UpperCamelCase__ , UpperCamelCase__ ) # Testing spaces after special tokens __UpperCamelCase ='''<mask>''' tokenizer.add_special_tokens( {'''mask_token''': AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ )} ) # mask token has a left space __UpperCamelCase =tokenizer.convert_tokens_to_ids(UpperCamelCase__ ) __UpperCamelCase ='''Encode <mask> sequence''' __UpperCamelCase ='''Encode <mask>sequence''' __UpperCamelCase =tokenizer.encode(UpperCamelCase__ ) __UpperCamelCase =encoded.index(UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) __UpperCamelCase =tokenizer.encode(UpperCamelCase__ ) __UpperCamelCase =encoded.index(UpperCamelCase__ ) __UpperCamelCase =tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(UpperCamelCase__ , UpperCamelCase__ ) def UpperCAmelCase_ ( self : int ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __UpperCamelCase =self.rust_tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__ ) __UpperCamelCase =self.tokenizer_class.from_pretrained(UpperCamelCase__ , **UpperCamelCase__ ) __UpperCamelCase ='''A, <mask> AllenNLP sentence.''' __UpperCamelCase =tokenizer_r.encode_plus(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ ) __UpperCamelCase =tokenizer_p.encode_plus(UpperCamelCase__ , add_special_tokens=UpperCamelCase__ , return_token_type_ids=UpperCamelCase__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['''token_type_ids'''] ) , sum(tokens_p['''token_type_ids'''] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['''attention_mask'''] ) / len(tokens_r['''attention_mask'''] ) , sum(tokens_p['''attention_mask'''] ) / len(tokens_p['''attention_mask'''] ) , ) __UpperCamelCase =tokenizer_r.convert_ids_to_tokens(tokens_r['''input_ids'''] ) __UpperCamelCase =tokenizer_p.convert_ids_to_tokens(tokens_p['''input_ids'''] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['''input_ids'''] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r['''input_ids'''] , [0, 250, 6, 50264, 3823, 487, 21992, 3645, 4, 2] ) self.assertSequenceEqual( UpperCamelCase__ , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) self.assertSequenceEqual( UpperCamelCase__ , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) __UpperCamelCase =json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state['''add_prefix_space'''] , UpperCamelCase__ ) self.assertEqual(post_processor_state['''add_prefix_space'''] , UpperCamelCase__ ) self.assertEqual(post_processor_state['''trim_offsets'''] , UpperCamelCase__ ) def UpperCAmelCase_ ( self : List[Any] ) -> int: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __UpperCamelCase ='''hello''' # `hello` is a token in the vocabulary of `pretrained_name` __UpperCamelCase =f"""{text_of_1_token} {text_of_1_token}""" __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ) + 1, len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ) + 1, len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ), len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (len(UpperCamelCase__ ), len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =f""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCamelCase__ ) + 1, 1 + len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCamelCase__ ), 1 + len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , ) __UpperCamelCase =self.rust_tokenizer_class.from_pretrained( UpperCamelCase__ , use_fast=UpperCamelCase__ , add_prefix_space=UpperCamelCase__ , trim_offsets=UpperCamelCase__ ) __UpperCamelCase =tokenizer_r(UpperCamelCase__ , return_offsets_mapping=UpperCamelCase__ , add_special_tokens=UpperCamelCase__ ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(UpperCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(UpperCamelCase__ ), 1 + len(UpperCamelCase__ ) + 1 + len(UpperCamelCase__ )) , )
85
0
"""simple docstring""" import argparse import os import re import packaging.version _UpperCamelCase: str = 'examples/' _UpperCamelCase: List[str] = { 'examples': (re.compile(R'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(R'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(R'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), R'\1version="VERSION",'), 'doc': (re.compile(R'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } _UpperCamelCase: Any = { 'init': 'src/diffusers/__init__.py', 'setup': 'setup.py', } _UpperCamelCase: Union[str, Any] = 'README.md' def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Dict: '''simple docstring''' with open(lowerCamelCase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: lowercase : List[str] = f.read() lowercase , lowercase : int = REPLACE_PATTERNS[pattern] lowercase : str = replace.replace('VERSION' , lowerCamelCase__ ) lowercase : Dict = re_pattern.sub(lowerCamelCase__ , lowerCamelCase__ ) with open(lowerCamelCase__ , 'w' , encoding='utf-8' , newline='\n' ) as f: f.write(lowerCamelCase__ ) def lowercase__ ( _UpperCAmelCase ) -> Dict: '''simple docstring''' for folder, directories, fnames in os.walk(lowerCamelCase__ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('research_projects' ) if "legacy" in directories: directories.remove('legacy' ) for fname in fnames: if fname.endswith('.py' ): update_version_in_file(os.path.join(lowerCamelCase__ , lowerCamelCase__ ) , lowerCamelCase__ , pattern='examples' ) def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase=False ) -> str: '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) if not patch: update_version_in_examples(lowerCamelCase__ ) def lowercase__ ( ) -> Any: '''simple docstring''' lowercase : str = '🤗 Transformers currently provides the following architectures' lowercase : Optional[int] = '1. Want to contribute a new model?' with open(lowerCamelCase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: lowercase : Union[str, Any] = f.readlines() # Find the start of the list. lowercase : str = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowercase : Union[str, Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('1.' ): lowercase : str = lines[index].replace( 'https://huggingface.co/docs/diffusers/main/model_doc' , 'https://huggingface.co/docs/diffusers/model_doc' , ) index += 1 with open(lowerCamelCase__ , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(lowerCamelCase__ ) def lowercase__ ( ) -> Optional[int]: '''simple docstring''' with open(REPLACE_FILES['init'] , 'r' ) as f: lowercase : Union[str, Any] = f.read() lowercase : Optional[Any] = REPLACE_PATTERNS['init'][0].search(lowerCamelCase__ ).groups()[0] return packaging.version.parse(lowerCamelCase__ ) def lowercase__ ( _UpperCAmelCase=False ) -> List[Any]: '''simple docstring''' lowercase : List[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('Can\'t create a patch version from the dev branch, checkout a released version!' ) if default_version.is_devrelease: lowercase : Union[str, Any] = default_version.base_version elif patch: lowercase : Dict = f'''{default_version.major}.{default_version.minor}.{default_version.micro + 1}''' else: lowercase : Tuple = f'''{default_version.major}.{default_version.minor + 1}.0''' # Now let's ask nicely if that's the right one. lowercase : Tuple = input(f'''Which version are you releasing? [{default_version}]''' ) if len(lowerCamelCase__ ) == 0: lowercase : Dict = default_version print(f'''Updating version to {version}.''' ) global_version_update(lowerCamelCase__ , patch=lowerCamelCase__ ) def lowercase__ ( ) -> Optional[int]: '''simple docstring''' lowercase : int = get_version() lowercase : Optional[Any] = f'''{current_version.major}.{current_version.minor + 1}.0.dev0''' lowercase : Any = current_version.base_version # Check with the user we got that right. lowercase : int = input(f'''Which version are we developing now? [{dev_version}]''' ) if len(lowerCamelCase__ ) == 0: lowercase : str = dev_version print(f'''Updating version to {version}.''' ) global_version_update(lowerCamelCase__ ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": _UpperCamelCase: Optional[Any] = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') _UpperCamelCase: Any = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
255
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: __A =None __A =logging.get_logger(__name__) __A ={'''vocab_file''': '''sentencepiece.bpe.model''', '''tokenizer_file''': '''tokenizer.json'''} __A ={ '''vocab_file''': { '''facebook/mbart-large-en-ro''': ( '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model''' ), '''facebook/mbart-large-cc25''': ( '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model''' ), }, '''tokenizer_file''': { '''facebook/mbart-large-en-ro''': '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json''', '''facebook/mbart-large-cc25''': '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json''', }, } __A ={ '''facebook/mbart-large-en-ro''': 1_0_2_4, '''facebook/mbart-large-cc25''': 1_0_2_4, } # fmt: off __A =['''ar_AR''', '''cs_CZ''', '''de_DE''', '''en_XX''', '''es_XX''', '''et_EE''', '''fi_FI''', '''fr_XX''', '''gu_IN''', '''hi_IN''', '''it_IT''', '''ja_XX''', '''kk_KZ''', '''ko_KR''', '''lt_LT''', '''lv_LV''', '''my_MM''', '''ne_NP''', '''nl_XX''', '''ro_RO''', '''ru_RU''', '''si_LK''', '''tr_TR''', '''vi_VN''', '''zh_CN'''] class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = ['input_ids', 'attention_mask'] lowerCAmelCase__ = MBartTokenizer lowerCAmelCase__ = [] lowerCAmelCase__ = [] def __init__( self , lowercase=None , lowercase=None , lowercase="<s>" , lowercase="</s>" , lowercase="</s>" , lowercase="<s>" , lowercase="<unk>" , lowercase="<pad>" , lowercase="<mask>" , lowercase=None , lowercase=None , lowercase=None , **lowercase , ) -> Dict: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase_ = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase ) if isinstance(lowercase , lowercase ) else mask_token super().__init__( vocab_file=lowercase , tokenizer_file=lowercase , bos_token=lowercase , eos_token=lowercase , sep_token=lowercase , cls_token=lowercase , unk_token=lowercase , pad_token=lowercase , mask_token=lowercase , src_lang=lowercase , tgt_lang=lowercase , additional_special_tokens=lowercase , **lowercase , ) lowerCamelCase_ = vocab_file lowerCamelCase_ = False if not self.vocab_file else True lowerCamelCase_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) lowerCamelCase_ = { lang_code: self.convert_tokens_to_ids(lowercase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } lowerCamelCase_ = src_lang if src_lang is not None else "en_XX" lowerCamelCase_ = self.convert_tokens_to_ids(self._src_lang ) lowerCamelCase_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def SCREAMING_SNAKE_CASE_( self ) -> str: return self._src_lang @src_lang.setter def SCREAMING_SNAKE_CASE_( self , lowercase ) -> None: lowerCamelCase_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , **lowercase ) -> List[Any]: if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) lowerCamelCase_ = src_lang lowerCamelCase_ = self(lowercase , add_special_tokens=lowercase , return_tensors=lowercase , **lowercase ) lowerCamelCase_ = self.convert_tokens_to_ids(lowercase ) lowerCamelCase_ = tgt_lang_id return inputs def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = "en_XX" , lowercase = None , lowercase = "ro_RO" , **lowercase , ) -> BatchEncoding: lowerCamelCase_ = src_lang lowerCamelCase_ = tgt_lang return super().prepare_seqaseq_batch(lowercase , lowercase , **lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: return self.set_src_lang_special_tokens(self.src_lang ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> None: lowerCamelCase_ = self.convert_tokens_to_ids(lowercase ) lowerCamelCase_ = [] lowerCamelCase_ = [self.eos_token_id, self.cur_lang_code] lowerCamelCase_ = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase_ = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase_ = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> None: lowerCamelCase_ = self.convert_tokens_to_ids(lowercase ) lowerCamelCase_ = [] lowerCamelCase_ = [self.eos_token_id, self.cur_lang_code] lowerCamelCase_ = self.convert_ids_to_tokens(self.prefix_tokens ) lowerCamelCase_ = self.convert_ids_to_tokens(self.suffix_tokens ) lowerCamelCase_ = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowercase ): logger.error(f'Vocabulary path ({save_directory}) should be a directory.' ) return lowerCamelCase_ = os.path.join( lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase ): copyfile(self.vocab_file , lowercase ) return (out_vocab_file,)
19
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase_ = { "configuration_resnet": ["RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "ResNetConfig", "ResNetOnnxConfig"] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ "RESNET_PRETRAINED_MODEL_ARCHIVE_LIST", "ResNetForImageClassification", "ResNetModel", "ResNetPreTrainedModel", "ResNetBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ "TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFResNetForImageClassification", "TFResNetModel", "TFResNetPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ "FlaxResNetForImageClassification", "FlaxResNetModel", "FlaxResNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()["__file__"], _import_structure)
332
'''simple docstring''' import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def __magic_name__ ( A ) -> Tuple: snake_case = [] embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight''', F'''stage{idx}.patch_embed.proj.weight''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias''', F'''stage{idx}.patch_embed.proj.bias''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight''', F'''stage{idx}.patch_embed.norm.weight''', ) ) embed.append( ( F'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias''', F'''stage{idx}.patch_embed.norm.bias''', ) ) return embed def __magic_name__ ( A , A ) -> Optional[int]: snake_case = [] attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked''', F'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_q.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_q.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_k.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_k.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj_v.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj_v.bias''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight''', F'''stage{idx}.blocks.{cnt}.attn.proj.weight''', ) ) attention_weights.append( ( F'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias''', F'''stage{idx}.blocks.{cnt}.attn.proj.bias''', ) ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight''', F'''stage{idx}.blocks.{cnt}.mlp.fc1.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias''', F'''stage{idx}.blocks.{cnt}.mlp.fc1.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight''', F'''stage{idx}.blocks.{cnt}.mlp.fc2.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias''', F'''stage{idx}.blocks.{cnt}.mlp.fc2.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight''', F'''stage{idx}.blocks.{cnt}.norm1.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias''', F'''stage{idx}.blocks.{cnt}.norm1.bias''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight''', F'''stage{idx}.blocks.{cnt}.norm2.weight''') ) attention_weights.append( (F'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias''', F'''stage{idx}.blocks.{cnt}.norm2.bias''') ) return attention_weights def __magic_name__ ( A ) -> List[Any]: snake_case = [] token.append((F'''cvt.encoder.stages.{idx}.cls_token''', 'stage2.cls_token') ) return token def __magic_name__ ( ) -> Dict: snake_case = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def __magic_name__ ( A , A , A , A ) -> int: snake_case = 'imagenet-1k-id2label.json' snake_case = 1_0_0_0 snake_case = 'huggingface/label-files' snake_case = num_labels snake_case = json.load(open(cached_download(hf_hub_url(A , A , repo_type='dataset' ) ) , 'r' ) ) snake_case = {int(A ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} snake_case = snake_case = CvtConfig(num_labels=A , idalabel=A , labelaid=A ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": snake_case = [1, 2, 1_0] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": snake_case = [1, 4, 1_6] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: snake_case = [2, 2, 2_0] snake_case = [3, 1_2, 1_6] snake_case = [1_9_2, 7_6_8, 1_0_2_4] snake_case = CvtForImageClassification(A ) snake_case = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) snake_case = image_size snake_case = torch.load(A , map_location=torch.device('cpu' ) ) snake_case = OrderedDict() snake_case = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: snake_case = list_of_state_dict + cls_token(A ) snake_case = list_of_state_dict + embeddings(A ) for cnt in range(config.depth[idx] ): snake_case = list_of_state_dict + attention(A , A ) snake_case = list_of_state_dict + final() for gg in list_of_state_dict: print(A ) for i in range(len(A ) ): snake_case = original_weights[list_of_state_dict[i][1]] model.load_state_dict(A ) model.save_pretrained(A ) image_processor.save_pretrained(A ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": lowerCAmelCase_ = argparse.ArgumentParser() parser.add_argument( "--cvt_model", default="cvt-w24", type=str, help="Name of the cvt model you'd like to convert.", ) parser.add_argument( "--image_size", default=3_8_4, type=int, help="Input Image Size", ) parser.add_argument( "--cvt_file_name", default=r"cvtmodels\CvT-w24-384x384-IN-22k.pth", type=str, help="Input Image Size", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) lowerCAmelCase_ = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
332
1
'''simple docstring''' import argparse import ast import logging import os import sys import pandas as pd import torch from tqdm import tqdm from transformers import BartForConditionalGeneration, RagRetriever, RagSequenceForGeneration, RagTokenForGeneration from transformers import logging as transformers_logging sys.path.append(os.path.join(os.getcwd())) # noqa: E402 # isort:skip from utils_rag import exact_match_score, fa_score # noqa: E402 # isort:skip UpperCamelCase__: Union[str, Any] = logging.getLogger(__name__) logging.basicConfig(level=logging.INFO) transformers_logging.set_verbosity_info() def snake_case_ ( _lowerCAmelCase : str ) -> Optional[int]: if "token" in model_name_or_path: return "rag_token" if "sequence" in model_name_or_path: return "rag_sequence" if "bart" in model_name_or_path: return "bart" return None def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str ) -> Dict: return max(metric_fn(_lowerCAmelCase , _lowerCAmelCase ) for gt in ground_truths ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> List[str]: UpperCAmelCase : str = [line.strip() for line in open(_lowerCAmelCase , '''r''' ).readlines()] UpperCAmelCase : Optional[Any] = [] if args.gold_data_mode == "qa": UpperCAmelCase : Optional[int] = pd.read_csv(_lowerCAmelCase , sep='''\t''' , header=_lowerCAmelCase ) for answer_list in data[1]: UpperCAmelCase : int = ast.literal_eval(_lowerCAmelCase ) answers.append(_lowerCAmelCase ) else: UpperCAmelCase : int = [line.strip() for line in open(_lowerCAmelCase , '''r''' ).readlines()] UpperCAmelCase : Optional[Any] = [[reference] for reference in references] UpperCAmelCase : Optional[int] = 0 for prediction, ground_truths in zip(_lowerCAmelCase , _lowerCAmelCase ): total += 1 em += metric_max_over_ground_truths(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) fa += metric_max_over_ground_truths(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = 1_0_0.0 * em / total UpperCAmelCase : Any = 1_0_0.0 * fa / total logger.info(f"""F1: {fa:.2f}""" ) logger.info(f"""EM: {em:.2f}""" ) def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] ) -> Optional[Any]: UpperCAmelCase : Union[str, Any] = args.k UpperCAmelCase : Tuple = [line.strip() for line in open(_lowerCAmelCase , '''r''' ).readlines()] UpperCAmelCase : Tuple = [line.strip() for line in open(_lowerCAmelCase , '''r''' ).readlines()] UpperCAmelCase : Any = 0 for hypo, reference in zip(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : Dict = set(hypo.split('''\t''' )[:k] ) UpperCAmelCase : Optional[Any] = set(reference.split('''\t''' ) ) total += 1 em += len(hypo_provenance & ref_provenance ) / k UpperCAmelCase : str = 1_0_0.0 * em / total logger.info(f"""Precision@{k}: {em: .2f}""" ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : int , _lowerCAmelCase : Dict ) -> Optional[int]: def strip_title(_lowerCAmelCase : Optional[int] ): if title.startswith('''"''' ): UpperCAmelCase : Tuple = title[1:] if title.endswith('''"''' ): UpperCAmelCase : Optional[Any] = title[:-1] return title UpperCAmelCase : Union[str, Any] = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( _lowerCAmelCase , return_tensors='''pt''' , padding=_lowerCAmelCase , truncation=_lowerCAmelCase , )['''input_ids'''].to(args.device ) UpperCAmelCase : str = rag_model.rag.question_encoder(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = question_enc_outputs[0] UpperCAmelCase : Any = rag_model.retriever( _lowerCAmelCase , question_enc_pool_output.cpu().detach().to(torch.floataa ).numpy() , prefix=rag_model.rag.generator.config.prefix , n_docs=rag_model.config.n_docs , return_tensors='''pt''' , ) UpperCAmelCase : int = rag_model.retriever.index.get_doc_dicts(result.doc_ids ) UpperCAmelCase : int = [] for docs in all_docs: UpperCAmelCase : Optional[int] = [strip_title(_lowerCAmelCase ) for title in docs['''title''']] provenance_strings.append('''\t'''.join(_lowerCAmelCase ) ) return provenance_strings def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : Dict ) -> List[str]: with torch.no_grad(): UpperCAmelCase : Any = rag_model.retriever.question_encoder_tokenizer.batch_encode_plus( _lowerCAmelCase , return_tensors='''pt''' , padding=_lowerCAmelCase , truncation=_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = inputs_dict.input_ids.to(args.device ) UpperCAmelCase : Union[str, Any] = inputs_dict.attention_mask.to(args.device ) UpperCAmelCase : Optional[int] = rag_model.generate( # rag_model overwrites generate _lowerCAmelCase , attention_mask=_lowerCAmelCase , num_beams=args.num_beams , min_length=args.min_length , max_length=args.max_length , early_stopping=_lowerCAmelCase , num_return_sequences=1 , bad_words_ids=[[0, 0]] , ) UpperCAmelCase : str = rag_model.retriever.generator_tokenizer.batch_decode(_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) if args.print_predictions: for q, a in zip(_lowerCAmelCase , _lowerCAmelCase ): logger.info('''Q: {} - A: {}'''.format(_lowerCAmelCase , _lowerCAmelCase ) ) return answers def snake_case_ ( ) -> List[Any]: UpperCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( '''--model_type''' , choices=['''rag_sequence''', '''rag_token''', '''bart'''] , type=_lowerCAmelCase , help=( '''RAG model type: rag_sequence, rag_token or bart, if none specified, the type is inferred from the''' ''' model_name_or_path''' ) , ) parser.add_argument( '''--index_name''' , default=_lowerCAmelCase , choices=['''exact''', '''compressed''', '''legacy'''] , type=_lowerCAmelCase , help='''RAG model retriever type''' , ) parser.add_argument( '''--index_path''' , default=_lowerCAmelCase , type=_lowerCAmelCase , help='''Path to the retrieval index''' , ) parser.add_argument('''--n_docs''' , default=5 , type=_lowerCAmelCase , help='''Number of retrieved docs''' ) parser.add_argument( '''--model_name_or_path''' , default=_lowerCAmelCase , type=_lowerCAmelCase , required=_lowerCAmelCase , help='''Path to pretrained checkpoints or model identifier from huggingface.co/models''' , ) parser.add_argument( '''--eval_mode''' , choices=['''e2e''', '''retrieval'''] , default='''e2e''' , type=_lowerCAmelCase , help=( '''Evaluation mode, e2e calculates exact match and F1 of the downstream task, retrieval calculates''' ''' precision@k.''' ) , ) parser.add_argument('''--k''' , default=1 , type=_lowerCAmelCase , help='''k for the precision@k calculation''' ) parser.add_argument( '''--evaluation_set''' , default=_lowerCAmelCase , type=_lowerCAmelCase , required=_lowerCAmelCase , help='''Path to a file containing evaluation samples''' , ) parser.add_argument( '''--gold_data_path''' , default=_lowerCAmelCase , type=_lowerCAmelCase , required=_lowerCAmelCase , help='''Path to a tab-separated file with gold samples''' , ) parser.add_argument( '''--gold_data_mode''' , default='''qa''' , type=_lowerCAmelCase , choices=['''qa''', '''ans'''] , help=( '''Format of the gold data file''' '''qa - a single line in the following format: question [tab] answer_list''' '''ans - a single line of the gold file contains the expected answer string''' ) , ) parser.add_argument( '''--predictions_path''' , type=_lowerCAmelCase , default='''predictions.txt''' , help='''Name of the predictions file, to be stored in the checkpoints directory''' , ) parser.add_argument( '''--eval_all_checkpoints''' , action='''store_true''' , help='''Evaluate all checkpoints starting with the same prefix as model_name ending and ending with step number''' , ) parser.add_argument( '''--eval_batch_size''' , default=8 , type=_lowerCAmelCase , help='''Batch size per GPU/CPU for evaluation.''' , ) parser.add_argument( '''--recalculate''' , help='''Recalculate predictions even if the prediction file exists''' , action='''store_true''' , ) parser.add_argument( '''--num_beams''' , default=4 , type=_lowerCAmelCase , help='''Number of beams to be used when generating answers''' , ) parser.add_argument('''--min_length''' , default=1 , type=_lowerCAmelCase , help='''Min length of the generated answers''' ) parser.add_argument('''--max_length''' , default=50 , type=_lowerCAmelCase , help='''Max length of the generated answers''' ) parser.add_argument( '''--print_predictions''' , action='''store_true''' , help='''If True, prints predictions while evaluating.''' , ) parser.add_argument( '''--print_docs''' , action='''store_true''' , help='''If True, prints docs retried while generating.''' , ) UpperCAmelCase : Any = parser.parse_args() UpperCAmelCase : str = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) return args def snake_case_ ( _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : int = {} if args.model_type is None: UpperCAmelCase : Union[str, Any] = infer_model_type(args.model_name_or_path ) assert args.model_type is not None if args.model_type.startswith('''rag''' ): UpperCAmelCase : Optional[Any] = RagTokenForGeneration if args.model_type == '''rag_token''' else RagSequenceForGeneration UpperCAmelCase : Dict = args.n_docs if args.index_name is not None: UpperCAmelCase : Dict = args.index_name if args.index_path is not None: UpperCAmelCase : str = args.index_path else: UpperCAmelCase : int = BartForConditionalGeneration UpperCAmelCase : Optional[Any] = ( [f.path for f in os.scandir(args.model_name_or_path ) if f.is_dir()] if args.eval_all_checkpoints else [args.model_name_or_path] ) logger.info('''Evaluate the following checkpoints: %s''' , _lowerCAmelCase ) UpperCAmelCase : str = get_scores if args.eval_mode == '''e2e''' else get_precision_at_k UpperCAmelCase : Any = evaluate_batch_eae if args.eval_mode == '''e2e''' else evaluate_batch_retrieval for checkpoint in checkpoints: if os.path.exists(args.predictions_path ) and (not args.recalculate): logger.info('''Calculating metrics based on an existing predictions file: {}'''.format(args.predictions_path ) ) score_fn(_lowerCAmelCase , args.predictions_path , args.gold_data_path ) continue logger.info('''***** Running evaluation for {} *****'''.format(_lowerCAmelCase ) ) logger.info(''' Batch size = %d''' , args.eval_batch_size ) logger.info(''' Predictions will be stored under {}'''.format(args.predictions_path ) ) if args.model_type.startswith('''rag''' ): UpperCAmelCase : int = RagRetriever.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = model_class.from_pretrained(_lowerCAmelCase , retriever=_lowerCAmelCase , **_lowerCAmelCase ) model.retriever.init_retrieval() else: UpperCAmelCase : Union[str, Any] = model_class.from_pretrained(_lowerCAmelCase , **_lowerCAmelCase ) model.to(args.device ) with open(args.evaluation_set , '''r''' ) as eval_file, open(args.predictions_path , '''w''' ) as preds_file: UpperCAmelCase : List[str] = [] for line in tqdm(_lowerCAmelCase ): questions.append(line.strip() ) if len(_lowerCAmelCase ) == args.eval_batch_size: UpperCAmelCase : Any = evaluate_batch_fn(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) preds_file.write('''\n'''.join(_lowerCAmelCase ) + '''\n''' ) preds_file.flush() UpperCAmelCase : str = [] if len(_lowerCAmelCase ) > 0: UpperCAmelCase : Optional[int] = evaluate_batch_fn(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) preds_file.write('''\n'''.join(_lowerCAmelCase ) ) preds_file.flush() score_fn(_lowerCAmelCase , args.predictions_path , args.gold_data_path ) if __name__ == "__main__": UpperCamelCase__: List[str] = get_args() main(args)
23
def A_ ( _lowerCAmelCase ) -> str: UpperCamelCase : List[Any] = "" for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def A_ ( _lowerCAmelCase ) -> dict[str, str]: UpperCamelCase : Optional[Any] = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key UpperCamelCase : Tuple = remove_duplicates(key.upper() ) UpperCamelCase : int = len(_lowerCAmelCase ) # First fill cipher with key characters UpperCamelCase : int = {alphabet[i]: char for i, char in enumerate(_lowerCAmelCase )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(_lowerCAmelCase ) , 26 ): UpperCamelCase : Optional[Any] = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 UpperCamelCase : List[str] = alphabet[i - offset] UpperCamelCase : List[Any] = char return cipher_alphabet def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> str: return "".join(cipher_map.get(_lowerCAmelCase , _lowerCAmelCase ) for ch in message.upper() ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> str: UpperCamelCase : Union[str, Any] = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(_lowerCAmelCase , _lowerCAmelCase ) for ch in message.upper() ) def A_ ( ) -> None: UpperCamelCase : int = input("Enter message to encode or decode: " ).strip() UpperCamelCase : str = input("Enter keyword: " ).strip() UpperCamelCase : Union[str, Any] = input("Encipher or decipher? E/D:" ).strip()[0].lower() try: UpperCamelCase : List[str] = {"e": encipher, "d": decipher}[option] except KeyError: raise KeyError("invalid input option" ) UpperCamelCase : str = create_cipher_map(_lowerCAmelCase ) print(func(_lowerCAmelCase , _lowerCAmelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
52
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" @slow def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" __UpperCAmelCase : Optional[Any] = XLMRobertaModel.from_pretrained("xlm-roberta-base" ) __UpperCAmelCase : Optional[int] = torch.tensor([[0, 581, 10_269, 83, 99_942, 136, 60_742, 23, 70, 80_583, 18_276, 2]] ) # The dog is cute and lives in the garden house __UpperCAmelCase : int = torch.Size((1, 12, 768) ) # batch_size, sequence_length, embedding_vector_dim __UpperCAmelCase : Tuple = torch.tensor( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __UpperCAmelCase : List[str] = model(UpperCAmelCase_ )["last_hidden_state"].detach() self.assertEqual(output.shape , UpperCAmelCase_ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , UpperCAmelCase_ , atol=1e-3 ) ) @slow def lowerCamelCase_ ( self : Optional[int] ): """simple docstring""" __UpperCAmelCase : int = XLMRobertaModel.from_pretrained("xlm-roberta-large" ) __UpperCAmelCase : Tuple = torch.tensor([[0, 581, 10_269, 83, 99_942, 136, 60_742, 23, 70, 80_583, 18_276, 2]] ) # The dog is cute and lives in the garden house __UpperCAmelCase : Optional[int] = torch.Size((1, 12, 1_024) ) # batch_size, sequence_length, embedding_vector_dim __UpperCAmelCase : Any = torch.tensor( [[-0.0699, -0.0318, 0.0705, -0.1241, 0.0999, -0.0520, 0.1004, -0.1838, -0.4704, 0.1437, 0.0821, 0.0126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): __UpperCAmelCase : Optional[Any] = model(UpperCAmelCase_ )["last_hidden_state"].detach() self.assertEqual(output.shape , UpperCAmelCase_ ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , UpperCAmelCase_ , atol=1e-3 ) )
37
'''simple docstring''' import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 lowerCAmelCase__ : Dict = get_tests_dir("fixtures/dummy-config.json") class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" def lowerCamelCase_ ( self : str ): """simple docstring""" __UpperCAmelCase : str = 0 def lowerCamelCase_ ( self : Tuple ): """simple docstring""" self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec("transformers.models.auto" ) ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" __UpperCAmelCase : List[str] = AutoConfig.from_pretrained("bert-base-uncased" ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" __UpperCAmelCase : Tuple = AutoConfig.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Any ): """simple docstring""" __UpperCAmelCase : Any = AutoConfig.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = AutoConfig.for_model("roberta" ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. __UpperCAmelCase : int = os.path.join(UpperCAmelCase_ , "fake-roberta" ) os.makedirs(UpperCAmelCase_ , exist_ok=UpperCAmelCase_ ) with open(os.path.join(UpperCAmelCase_ , "config.json" ) , "w" ) as f: f.write(json.dumps({} ) ) __UpperCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(UpperCAmelCase_ ) self.assertEqual(type(UpperCAmelCase_ ) , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" try: AutoConfig.register("custom" , UpperCAmelCase_ ) # Wrong model type will raise an error with self.assertRaises(UpperCAmelCase_ ): AutoConfig.register("model" , UpperCAmelCase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(UpperCAmelCase_ ): AutoConfig.register("bert" , UpperCAmelCase_ ) # Now that the config is registered, it can be used as any other config with the auto-API __UpperCAmelCase : List[str] = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(UpperCAmelCase_ ) __UpperCAmelCase : Dict = AutoConfig.from_pretrained(UpperCAmelCase_ ) self.assertIsInstance(UpperCAmelCase_ , UpperCAmelCase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def lowerCamelCase_ ( self : Union[str, Any] ): """simple docstring""" with self.assertRaisesRegex( UpperCAmelCase_ , "bert-base is not a local folder and is not a valid model identifier" ): __UpperCAmelCase : Union[str, Any] = AutoConfig.from_pretrained("bert-base" ) def lowerCamelCase_ ( self : int ): """simple docstring""" with self.assertRaisesRegex( UpperCAmelCase_ , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __UpperCAmelCase : int = AutoConfig.from_pretrained(UpperCAmelCase_ , revision="aaaaaa" ) def lowerCamelCase_ ( self : List[str] ): """simple docstring""" with self.assertRaisesRegex( UpperCAmelCase_ , "hf-internal-testing/no-config-test-repo does not appear to have a file named config.json." , ): __UpperCAmelCase : Tuple = AutoConfig.from_pretrained("hf-internal-testing/no-config-test-repo" ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(UpperCAmelCase_ ): __UpperCAmelCase : int = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) # If remote code is disabled, we can't load this config. with self.assertRaises(UpperCAmelCase_ ): __UpperCAmelCase : Any = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=UpperCAmelCase_ ) __UpperCAmelCase : Optional[Any] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=UpperCAmelCase_ ) self.assertEqual(config.__class__.__name__ , "NewModelConfig" ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(UpperCAmelCase_ ) __UpperCAmelCase : Union[str, Any] = AutoConfig.from_pretrained(UpperCAmelCase_ , trust_remote_code=UpperCAmelCase_ ) self.assertEqual(reloaded_config.__class__.__name__ , "NewModelConfig" ) def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( snake_case__ ): """simple docstring""" SCREAMING_SNAKE_CASE = '''new-model''' try: AutoConfig.register("new-model" , UpperCAmelCase_ ) # If remote code is not set, the default is to use local __UpperCAmelCase : Union[str, Any] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal" ) # If remote code is disabled, we load the local one. __UpperCAmelCase : int = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=UpperCAmelCase_ ) self.assertEqual(config.__class__.__name__ , "NewModelConfigLocal" ) # If remote is enabled, we load from the Hub __UpperCAmelCase : Tuple = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" , trust_remote_code=UpperCAmelCase_ ) self.assertEqual(config.__class__.__name__ , "NewModelConfig" ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
37
1
import inspect import logging import os import random import shutil import tempfile import unittest import pytest import torch from torch import nn from torch.utils.data import DataLoader, TensorDataset from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_cuda from accelerate.utils import ProjectConfiguration, set_seed a : Dict = logging.getLogger(__name__) def lowerCAmelCase_ (lowerCAmelCase__: List[str]=2 , lowerCAmelCase__: List[Any]=3 , lowerCAmelCase__: int=1_6 , lowerCAmelCase__: int = 1_0 , lowerCAmelCase__: int = 2 ): """simple docstring""" def get_dataset(lowerCAmelCase__: List[Any] ): UpperCAmelCase_: List[Any] = torch.randn(batch_size * n_batches , 1 ) return TensorDataset(lowerCAmelCase__ , a * x + b + 0.1 * torch.randn(batch_size * n_batches , 1 ) ) UpperCAmelCase_: Union[str, Any] = get_dataset(lowerCAmelCase__ ) UpperCAmelCase_: Optional[int] = get_dataset(lowerCAmelCase__ ) UpperCAmelCase_: str = DataLoader(lowerCAmelCase__ , shuffle=lowerCAmelCase__ , batch_size=lowerCAmelCase__ , num_workers=4 ) UpperCAmelCase_: Union[str, Any] = DataLoader(lowerCAmelCase__ , shuffle=lowerCAmelCase__ , batch_size=lowerCAmelCase__ , num_workers=4 ) return (train_dataloader, valid_dataloader) def lowerCAmelCase_ (lowerCAmelCase__: Optional[Any] , lowerCAmelCase__: Dict , lowerCAmelCase__: Union[str, Any] , lowerCAmelCase__: Tuple , lowerCAmelCase__: Union[str, Any] , lowerCAmelCase__: List[Any]=None ): """simple docstring""" UpperCAmelCase_: Union[str, Any] = [] for epoch in range(lowerCAmelCase__ ): # Train quickly model.train() for batch in dataloader: UpperCAmelCase_ , UpperCAmelCase_: str = batch UpperCAmelCase_: Tuple = model(lowerCAmelCase__ ) UpperCAmelCase_: Union[str, Any] = torch.nn.functional.mse_loss(lowerCAmelCase__ , lowerCAmelCase__ ) accelerator.backward(lowerCAmelCase__ ) optimizer.step() optimizer.zero_grad() rands.append(random.random() ) # Introduce some randomness if scheduler is not None: scheduler.step() return rands class _a ( nn.Module ): def __init__(self ) -> str: super().__init__() UpperCAmelCase_: Optional[Any] = nn.Parameter(torch.randn(1 ) ) UpperCAmelCase_: Optional[Any] = nn.Parameter(torch.randn(1 ) ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> List[Any]: return x * self.a + self.b class _a ( unittest.TestCase ): def __snake_case (self ) -> Union[str, Any]: with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase_: Dict = DummyModel() UpperCAmelCase_: List[str] = torch.optim.Adam(params=model.parameters(), lr=1E-3 ) UpperCAmelCase_ , UpperCAmelCase_: Dict = dummy_dataloaders() UpperCAmelCase_: Tuple = ProjectConfiguration(total_limit=1, project_dir=SCREAMING_SNAKE_CASE_, automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) # Train baseline UpperCAmelCase_: Union[str, Any] = Accelerator(project_config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_: str = accelerator.prepare( SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # Save initial accelerator.save_state() # Save second state accelerator.save_state() self.assertEqual(len(os.listdir(accelerator.project_dir ) ), 1 ) def __snake_case (self ) -> Optional[int]: with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase_: List[str] = DummyModel() UpperCAmelCase_: Any = torch.optim.Adam(params=model.parameters(), lr=1E-3 ) UpperCAmelCase_ , UpperCAmelCase_: Tuple = dummy_dataloaders() # Train baseline UpperCAmelCase_: Tuple = Accelerator() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_: Tuple = accelerator.prepare( SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # Save initial UpperCAmelCase_: Union[str, Any] = os.path.join(SCREAMING_SNAKE_CASE_, """initial""" ) accelerator.save_state(SCREAMING_SNAKE_CASE_ ) ((UpperCAmelCase_) , (UpperCAmelCase_)): Union[str, Any] = model.a.item(), model.b.item() UpperCAmelCase_: Union[str, Any] = optimizer.state_dict() UpperCAmelCase_: Tuple = train(3, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) ((UpperCAmelCase_) , (UpperCAmelCase_)): Union[str, Any] = model.a.item(), model.b.item() UpperCAmelCase_: Optional[Any] = optimizer.state_dict() # Train partially set_seed(42 ) UpperCAmelCase_: Any = DummyModel() UpperCAmelCase_: int = torch.optim.Adam(params=model.parameters(), lr=1E-3 ) UpperCAmelCase_ , UpperCAmelCase_: List[str] = dummy_dataloaders() UpperCAmelCase_: List[str] = Accelerator() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_: List[Any] = accelerator.prepare( SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) accelerator.load_state(SCREAMING_SNAKE_CASE_ ) ((UpperCAmelCase_) , (UpperCAmelCase_)): Optional[Any] = model.a.item(), model.b.item() UpperCAmelCase_: Optional[Any] = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[str] = train(2, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # Save everything UpperCAmelCase_: str = os.path.join(SCREAMING_SNAKE_CASE_, """checkpoint""" ) accelerator.save_state(SCREAMING_SNAKE_CASE_ ) # Load everything back in and make sure all states work accelerator.load_state(SCREAMING_SNAKE_CASE_ ) test_rands += train(1, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) ((UpperCAmelCase_) , (UpperCAmelCase_)): int = model.a.item(), model.b.item() UpperCAmelCase_: int = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> str: with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase_: Union[str, Any] = DummyModel() UpperCAmelCase_: str = torch.optim.Adam(params=model.parameters(), lr=1E-3 ) UpperCAmelCase_ , UpperCAmelCase_: Any = dummy_dataloaders() UpperCAmelCase_: Any = ProjectConfiguration(automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) # Train baseline UpperCAmelCase_: Optional[int] = Accelerator(project_dir=SCREAMING_SNAKE_CASE_, project_config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_: str = accelerator.prepare( SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # Save initial accelerator.save_state() ((UpperCAmelCase_) , (UpperCAmelCase_)): Any = model.a.item(), model.b.item() UpperCAmelCase_: Union[str, Any] = optimizer.state_dict() UpperCAmelCase_: Dict = train(3, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) ((UpperCAmelCase_) , (UpperCAmelCase_)): List[Any] = model.a.item(), model.b.item() UpperCAmelCase_: Any = optimizer.state_dict() # Train partially set_seed(42 ) UpperCAmelCase_: Any = DummyModel() UpperCAmelCase_: str = torch.optim.Adam(params=model.parameters(), lr=1E-3 ) UpperCAmelCase_ , UpperCAmelCase_: List[Any] = dummy_dataloaders() UpperCAmelCase_: Union[str, Any] = ProjectConfiguration(iteration=1, automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Union[str, Any] = Accelerator(project_dir=SCREAMING_SNAKE_CASE_, project_config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_: int = accelerator.prepare( SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) accelerator.load_state(os.path.join(SCREAMING_SNAKE_CASE_, """checkpoints""", """checkpoint_0""" ) ) ((UpperCAmelCase_) , (UpperCAmelCase_)): List[Any] = model.a.item(), model.b.item() UpperCAmelCase_: Optional[int] = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Optional[int] = train(2, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # Save everything accelerator.save_state() # Load everything back in and make sure all states work accelerator.load_state(os.path.join(SCREAMING_SNAKE_CASE_, """checkpoints""", """checkpoint_1""" ) ) test_rands += train(1, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) ((UpperCAmelCase_) , (UpperCAmelCase_)): Union[str, Any] = model.a.item(), model.b.item() UpperCAmelCase_: Tuple = optimizer.state_dict() self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) def __snake_case (self ) -> Tuple: UpperCAmelCase_: Optional[int] = torch.tensor([1, 2, 3] ) UpperCAmelCase_: Optional[int] = torch.tensor([2, 3, 4] ) UpperCAmelCase_: Union[str, Any] = DummyModel() UpperCAmelCase_: Optional[int] = torch.optim.Adam(net.parameters() ) UpperCAmelCase_: Dict = Accelerator() with self.assertRaises(SCREAMING_SNAKE_CASE_ ) as ve: accelerator.register_for_checkpointing(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: Union[str, Any] = str(ve.exception ) self.assertTrue("""Item at index 0""" in message ) self.assertTrue("""Item at index 1""" in message ) self.assertFalse("""Item at index 2""" in message ) self.assertFalse("""Item at index 3""" in message ) def __snake_case (self ) -> str: with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase_: Any = DummyModel() UpperCAmelCase_: Optional[int] = torch.optim.Adam(params=model.parameters(), lr=1E-3 ) UpperCAmelCase_: List[Any] = torch.optim.lr_scheduler.StepLR(SCREAMING_SNAKE_CASE_, step_size=1, gamma=0.9_9 ) UpperCAmelCase_ , UpperCAmelCase_: Any = dummy_dataloaders() UpperCAmelCase_: int = ProjectConfiguration(automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_ ) # Train baseline UpperCAmelCase_: List[Any] = Accelerator(project_dir=SCREAMING_SNAKE_CASE_, project_config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_: Any = accelerator.prepare( SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) # Save initial accelerator.save_state() UpperCAmelCase_: List[Any] = scheduler.state_dict() train(3, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) self.assertNotEqual(SCREAMING_SNAKE_CASE_, scheduler.state_dict() ) # Load everything back in and make sure all states work accelerator.load_state(os.path.join(SCREAMING_SNAKE_CASE_, """checkpoints""", """checkpoint_0""" ) ) self.assertEqual(SCREAMING_SNAKE_CASE_, scheduler.state_dict() ) def __snake_case (self ) -> str: with tempfile.TemporaryDirectory() as tmpdir: set_seed(42 ) UpperCAmelCase_: List[Any] = DummyModel() UpperCAmelCase_: Tuple = ProjectConfiguration(automatic_checkpoint_naming=SCREAMING_SNAKE_CASE_, total_limit=2 ) # Train baseline UpperCAmelCase_: List[str] = Accelerator(project_dir=SCREAMING_SNAKE_CASE_, project_config=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[str] = accelerator.prepare(SCREAMING_SNAKE_CASE_ ) # Save 3 states: for _ in range(11 ): accelerator.save_state() self.assertTrue(not os.path.exists(os.path.join(SCREAMING_SNAKE_CASE_, """checkpoints""", """checkpoint_0""" ) ) ) self.assertTrue(os.path.exists(os.path.join(SCREAMING_SNAKE_CASE_, """checkpoints""", """checkpoint_9""" ) ) ) self.assertTrue(os.path.exists(os.path.join(SCREAMING_SNAKE_CASE_, """checkpoints""", """checkpoint_10""" ) ) ) @require_cuda def __snake_case (self ) -> List[Any]: UpperCAmelCase_: List[str] = ["""torchrun""", f'--nproc_per_node={torch.cuda.device_count()}', inspect.getfile(self.__class__ )] execute_subprocess_async(SCREAMING_SNAKE_CASE_, env=os.environ.copy() ) if __name__ == "__main__": a : Any = '/tmp/accelerate/state_checkpointing' a : Tuple = DummyModel() a : Tuple = torch.optim.Adam(params=model.parameters(), lr=1E-3) a : Optional[Any] = torch.optim.lr_scheduler.StepLR(optimizer, step_size=1, gamma=0.9_9) a ,a : List[Any] = dummy_dataloaders() a : Dict = ProjectConfiguration(automatic_checkpoint_naming=True) # Train baseline a : str = Accelerator(project_dir=savedir, project_config=project_config, mixed_precision='no') if accelerator.process_index == 0: if os.path.exists(savedir): shutil.rmtree(savedir) os.makedirs(savedir) a ,a ,a ,a ,a : Optional[int] = accelerator.prepare( model, optimizer, train_dataloader, valid_dataloader, scheduler ) a ,a : Tuple = accelerator.prepare(model, optimizer) train(3, model, train_dataloader, optimizer, accelerator, scheduler) # Check that the intial optimizer is loaded on the GPU for group in optimizer.param_groups: a : List[Any] = group['params'][0].device break assert param_device.type == accelerator.device.type a : Dict = model.cpu() accelerator.wait_for_everyone() accelerator.save_state() accelerator.wait_for_everyone() # Check CPU state accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='cpu') for group in optimizer.param_groups: a : Union[str, Any] = group['params'][0].device break assert ( param_device.type == torch.device('cpu').type ), F"Loaded optimizer states did not match, expected to be loaded on the CPU but got {param_device}" # Check device state model.to(accelerator.device) accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='on_device') for group in optimizer.param_groups: a : Tuple = group['params'][0].device break assert ( param_device.type == accelerator.device.type ), F"Loaded optimizer states did not match, expected to be loaded on {accelerator.device} but got {param_device}" # Check error with pytest.raises(TypeError, match='Unsupported optimizer map location passed'): accelerator.load_state(os.path.join(savedir, 'checkpoints', 'checkpoint_0'), map_location='invalid') accelerator.wait_for_everyone() if accelerator.process_index == 0: shutil.rmtree(savedir) accelerator.wait_for_everyone()
147
a : Optional[int] = 9.8_0_6_6_5 def lowerCAmelCase_ (lowerCAmelCase__: float , lowerCAmelCase__: float , lowerCAmelCase__: float = g ): """simple docstring""" if fluid_density <= 0: raise ValueError("""Impossible fluid density""" ) if volume < 0: raise ValueError("""Impossible Object volume""" ) if gravity <= 0: raise ValueError("""Impossible Gravity""" ) return fluid_density * gravity * volume if __name__ == "__main__": import doctest # run doctest doctest.testmod()
147
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __A = { 'configuration_roberta_prelayernorm': [ 'ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RobertaPreLayerNormConfig', 'RobertaPreLayerNormOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ 'ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST', 'RobertaPreLayerNormForCausalLM', 'RobertaPreLayerNormForMaskedLM', 'RobertaPreLayerNormForMultipleChoice', 'RobertaPreLayerNormForQuestionAnswering', 'RobertaPreLayerNormForSequenceClassification', 'RobertaPreLayerNormForTokenClassification', 'RobertaPreLayerNormModel', 'RobertaPreLayerNormPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ 'TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRobertaPreLayerNormForCausalLM', 'TFRobertaPreLayerNormForMaskedLM', 'TFRobertaPreLayerNormForMultipleChoice', 'TFRobertaPreLayerNormForQuestionAnswering', 'TFRobertaPreLayerNormForSequenceClassification', 'TFRobertaPreLayerNormForTokenClassification', 'TFRobertaPreLayerNormMainLayer', 'TFRobertaPreLayerNormModel', 'TFRobertaPreLayerNormPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ 'FlaxRobertaPreLayerNormForCausalLM', 'FlaxRobertaPreLayerNormForMaskedLM', 'FlaxRobertaPreLayerNormForMultipleChoice', 'FlaxRobertaPreLayerNormForQuestionAnswering', 'FlaxRobertaPreLayerNormForSequenceClassification', 'FlaxRobertaPreLayerNormForTokenClassification', 'FlaxRobertaPreLayerNormModel', 'FlaxRobertaPreLayerNormPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
354
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 650, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "pytorch", "script": "run_ddp.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "tensorflow", "script": "run_tf_dist.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 600, "eval_accuracy": 0.6, "eval_loss": 0.7}, }, ] ) class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __A ( self: Any ) -> Optional[Any]: if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding='''utf-8''' , check=__A , ) assert hasattr(self , '''env''' ) def __A ( self: Any , __A: Optional[int] ) -> Union[str, Any]: _A = f"""{self.env.base_job_name}-{instance_count}-{"ddp" if "ddp" in self.script else "smd"}""" # distributed data settings _A = {'''smdistributed''': {'''dataparallel''': {'''enabled''': True}}} if self.script != '''run_ddp.py''' else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=__A , instance_count=__A , instance_type=self.instance_type , debugger_hook_config=__A , hyperparameters={**self.env.distributed_hyperparameters, '''model_name_or_path''': self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=__A , py_version='''py36''' , ) def __A ( self: Optional[int] , __A: Optional[int] ) -> List[Any]: TrainingJobAnalytics(__A ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(2,)] ) def __A ( self: Dict , __A: Tuple ) -> Union[str, Any]: # create estimator _A = self.create_estimator(__A ) # run training estimator.fit() # result dataframe _A = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _A = list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value'''] ) _A = list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value'''] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _A = ( Session().describe_training_job(estimator.latest_training_job.name ).get('''TrainingTimeInSeconds''' , 99_99_99 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy ) assert all(t <= self.results['''eval_loss'''] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , '''w''' ) as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} , __A )
75
0
"""simple docstring""" from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ : Any = logging.get_logger(__name__) UpperCAmelCase__ : str = { 'snap-research/efficientformer-l1-300': ( 'https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json' ), } class lowerCAmelCase_ (a__ ): """simple docstring""" __UpperCamelCase : List[str] = '''efficientformer''' def __init__(self , SCREAMING_SNAKE_CASE__ = [3, 2, 6, 4] , SCREAMING_SNAKE_CASE__ = [48, 96, 2_24, 4_48] , SCREAMING_SNAKE_CASE__ = [True, True, True, True] , SCREAMING_SNAKE_CASE__ = 4_48 , SCREAMING_SNAKE_CASE__ = 32 , SCREAMING_SNAKE_CASE__ = 4 , SCREAMING_SNAKE_CASE__ = 7 , SCREAMING_SNAKE_CASE__ = 5 , SCREAMING_SNAKE_CASE__ = 8 , SCREAMING_SNAKE_CASE__ = 4 , SCREAMING_SNAKE_CASE__ = 0.0 , SCREAMING_SNAKE_CASE__ = 16 , SCREAMING_SNAKE_CASE__ = 3 , SCREAMING_SNAKE_CASE__ = 3 , SCREAMING_SNAKE_CASE__ = 3 , SCREAMING_SNAKE_CASE__ = 2 , SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = 0.0 , SCREAMING_SNAKE_CASE__ = 1 , SCREAMING_SNAKE_CASE__ = True , SCREAMING_SNAKE_CASE__ = True , SCREAMING_SNAKE_CASE__ = 1E-5 , SCREAMING_SNAKE_CASE__ = "gelu" , SCREAMING_SNAKE_CASE__ = 0.02 , SCREAMING_SNAKE_CASE__ = 1E-12 , SCREAMING_SNAKE_CASE__ = 2_24 , SCREAMING_SNAKE_CASE__ = 1E-05 , **SCREAMING_SNAKE_CASE__ , ) -> None: """simple docstring""" super().__init__(**SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = hidden_act SCREAMING_SNAKE_CASE__ : Tuple = hidden_dropout_prob SCREAMING_SNAKE_CASE__ : Tuple = hidden_sizes SCREAMING_SNAKE_CASE__ : Tuple = num_hidden_layers SCREAMING_SNAKE_CASE__ : Any = num_attention_heads SCREAMING_SNAKE_CASE__ : Dict = initializer_range SCREAMING_SNAKE_CASE__ : int = layer_norm_eps SCREAMING_SNAKE_CASE__ : Dict = patch_size SCREAMING_SNAKE_CASE__ : Tuple = num_channels SCREAMING_SNAKE_CASE__ : Any = depths SCREAMING_SNAKE_CASE__ : Dict = mlp_expansion_ratio SCREAMING_SNAKE_CASE__ : int = downsamples SCREAMING_SNAKE_CASE__ : Optional[int] = dim SCREAMING_SNAKE_CASE__ : Tuple = key_dim SCREAMING_SNAKE_CASE__ : Tuple = attention_ratio SCREAMING_SNAKE_CASE__ : Union[str, Any] = resolution SCREAMING_SNAKE_CASE__ : List[str] = pool_size SCREAMING_SNAKE_CASE__ : Union[str, Any] = downsample_patch_size SCREAMING_SNAKE_CASE__ : Union[str, Any] = downsample_stride SCREAMING_SNAKE_CASE__ : int = downsample_pad SCREAMING_SNAKE_CASE__ : List[str] = drop_path_rate SCREAMING_SNAKE_CASE__ : Any = num_metaad_blocks SCREAMING_SNAKE_CASE__ : List[Any] = distillation SCREAMING_SNAKE_CASE__ : Tuple = use_layer_scale SCREAMING_SNAKE_CASE__ : List[Any] = layer_scale_init_value SCREAMING_SNAKE_CASE__ : List[Any] = image_size SCREAMING_SNAKE_CASE__ : str = batch_norm_eps
25
'''simple docstring''' from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def _lowerCAmelCase ( __snake_case : Tuple ) -> Dict: return {key.lstrip('-' ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def _lowerCAmelCase ( ) -> Tuple: __A : int = ArgumentParser( 'HuggingFace Datasets CLI tool' , usage='datasets-cli <command> [<args>]' , allow_abbrev=__snake_case ) __A : Optional[Any] = parser.add_subparsers(help='datasets-cli command helpers' ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(__snake_case ) EnvironmentCommand.register_subcommand(__snake_case ) TestCommand.register_subcommand(__snake_case ) RunBeamCommand.register_subcommand(__snake_case ) DummyDataCommand.register_subcommand(__snake_case ) # Parse args __A ,__A : Optional[Any] = parser.parse_known_args() if not hasattr(__snake_case , 'func' ): parser.print_help() exit(1 ) __A : Any = parse_unknown_args(__snake_case ) # Run __A : List[Any] = args.func(__snake_case , **__snake_case ) service.run() if __name__ == "__main__": main()
190
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' lowercase_ = ["image_processor", "tokenizer"] lowercase_ = "CLIPImageProcessor" lowercase_ = ("XLMRobertaTokenizer", "XLMRobertaTokenizerFast") def __init__( self : List[str] , _lowerCAmelCase : Optional[int]=None , _lowerCAmelCase : int=None , **_lowerCAmelCase : str ): SCREAMING_SNAKE_CASE_ = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , _lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ = kwargs.pop('feature_extractor' ) SCREAMING_SNAKE_CASE_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(_lowerCAmelCase , _lowerCAmelCase ) def __call__( self : List[str] , _lowerCAmelCase : Any=None , _lowerCAmelCase : Dict=None , _lowerCAmelCase : Union[str, Any]=None , **_lowerCAmelCase : str ): if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: SCREAMING_SNAKE_CASE_ = self.tokenizer(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if images is not None: SCREAMING_SNAKE_CASE_ = self.image_processor(_lowerCAmelCase , return_tensors=_lowerCAmelCase , **_lowerCAmelCase ) if text is not None and images is not None: SCREAMING_SNAKE_CASE_ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**_lowerCAmelCase ) , tensor_type=_lowerCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] , *_lowerCAmelCase : Dict , **_lowerCAmelCase : str ): return self.tokenizer.batch_decode(*_lowerCAmelCase , **_lowerCAmelCase ) def lowerCAmelCase_ ( self : str , *_lowerCAmelCase : List[Any] , **_lowerCAmelCase : List[Any] ): return self.tokenizer.decode(*_lowerCAmelCase , **_lowerCAmelCase ) @property def lowerCAmelCase_ ( self : List[Any] ): SCREAMING_SNAKE_CASE_ = self.tokenizer.model_input_names SCREAMING_SNAKE_CASE_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
355
from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def UpperCAmelCase_ ( __UpperCAmelCase : Optional[int] ) -> int: if not is_accelerate_available(): return method SCREAMING_SNAKE_CASE_ = version.parse(accelerate.__version__ ).base_version if version.parse(__UpperCAmelCase ) < version.parse('0.17.0' ): return method def wrapper(self : Optional[int] , *__UpperCAmelCase : Optional[Any] , **__UpperCAmelCase : Optional[Any] ): if hasattr(self , '_hf_hook' ) and hasattr(self._hf_hook , 'pre_forward' ): self._hf_hook.pre_forward(self ) return method(self , *__UpperCAmelCase , **__UpperCAmelCase ) return wrapper
210
0
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import subprocess from packaging.version import Version, parse from accelerate.commands.config.config_args import default_config_file, load_config_from_file lowerCAmelCase__ = '''Run commands across TPU VMs for initial setup before running `accelerate launch`.''' def _A ( A__=None ): """simple docstring""" if subparsers is not None: __lowercase = subparsers.add_parser('''tpu-config''' , description=_description ) else: __lowercase = argparse.ArgumentParser('''Accelerate tpu-config command''' , description=_description ) # Core arguments __lowercase = parser.add_argument_group( '''Config Arguments''' , '''Arguments that can be configured through `accelerate config`.''' ) config_args.add_argument( '''--config_file''' , type=A__ , default=A__ , help='''Path to the config file to use for accelerate.''' , ) config_args.add_argument( '''--tpu_name''' , default=A__ , help='''The name of the TPU to use. If not specified, will use the TPU specified in the config file.''' , ) config_args.add_argument( '''--tpu_zone''' , default=A__ , help='''The zone of the TPU to use. If not specified, will use the zone specified in the config file.''' , ) __lowercase = parser.add_argument_group('''TPU Arguments''' , '''Arguments for options ran inside the TPU.''' ) pod_args.add_argument( '''--use_alpha''' , action='''store_true''' , help='''Whether to use `gcloud alpha` when running the TPU training script instead of `gcloud`.''' , ) pod_args.add_argument( '''--command_file''' , default=A__ , help='''The path to the file containing the commands to run on the pod on startup.''' , ) pod_args.add_argument( '''--command''' , action='''append''' , nargs='''+''' , help='''A command to run on the pod. Can be passed multiple times.''' , ) pod_args.add_argument( '''--install_accelerate''' , action='''store_true''' , help='''Whether to install accelerate on the pod. Defaults to False.''' , ) pod_args.add_argument( '''--accelerate_version''' , default='''latest''' , help='''The version of accelerate to install on the pod. If not specified, will use the latest pypi version. Specify \'dev\' to install from GitHub.''' , ) pod_args.add_argument( '''--debug''' , action='''store_true''' , help='''If set, will print the command that would be run instead of running it.''' ) if subparsers is not None: parser.set_defaults(func=A__ ) return parser def _A ( A__ ): """simple docstring""" __lowercase = None # Get the default from the config file if it exists. if args.config_file is not None or os.path.isfile(A__ ): __lowercase = load_config_from_file(args.config_file ) if not args.command_file and defaults.command_file is not None and not args.command: __lowercase = defaults.command_file if not args.command and defaults.commands is not None: __lowercase = defaults.commands if not args.tpu_name: __lowercase = defaults.tpu_name if not args.tpu_zone: __lowercase = defaults.tpu_zone if args.accelerate_version == "dev": __lowercase = '''git+https://github.com/huggingface/accelerate.git''' elif args.accelerate_version == "latest": __lowercase = '''accelerate -U''' elif isinstance(parse(args.accelerate_version ) , A__ ): __lowercase = F"accelerate=={args.accelerate_version}" if not args.command_file and not args.command: raise ValueError('''You must specify either a command file or a command to run on the pod.''' ) if args.command_file: with open(args.command_file , '''r''' ) as f: __lowercase = [f.read().splitlines()] # To turn list of lists into list of strings if isinstance(args.command[0] , A__ ): __lowercase = [line for cmd in args.command for line in cmd] # Default to the shared folder and install accelerate __lowercase = ['''cd /usr/share'''] if args.install_accelerate: new_cmd += [F"pip install {args.accelerate_version}"] new_cmd += args.command __lowercase = '''; '''.join(A__ ) # Then send it to gcloud # Eventually try to use google-api-core to do this instead of subprocess __lowercase = ['''gcloud'''] if args.use_alpha: cmd += ["alpha"] cmd += [ "compute", "tpus", "tpu-vm", "ssh", args.tpu_name, "--zone", args.tpu_zone, "--command", args.command, "--worker", "all", ] if args.debug: print(F"Running {' '.join(A__ )}" ) return subprocess.run(A__ ) print('''Successfully setup pod.''' ) def _A ( ): """simple docstring""" __lowercase = tpu_command_parser() __lowercase = parser.parse_args() tpu_command_launcher(A__ )
104
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : Tuple = logging.get_logger(__name__) _lowerCAmelCase : Union[str, Any] = {"ctrl": "https://huggingface.co/ctrl/resolve/main/config.json"} class __magic_name__ ( lowerCAmelCase_ ): SCREAMING_SNAKE_CASE = 'ctrl' SCREAMING_SNAKE_CASE = ['past_key_values'] SCREAMING_SNAKE_CASE = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self , __snake_case=24_6534 , __snake_case=256 , __snake_case=1280 , __snake_case=8192 , __snake_case=48 , __snake_case=16 , __snake_case=0.1 , __snake_case=0.1 , __snake_case=1e-6 , __snake_case=0.02 , __snake_case=True , **__snake_case , ) -> Union[str, Any]: '''simple docstring''' __a =vocab_size __a =n_positions __a =n_embd __a =n_layer __a =n_head __a =dff __a =resid_pdrop __a =embd_pdrop __a =layer_norm_epsilon __a =initializer_range __a =use_cache super().__init__(**__snake_case )
218
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class __lowercase (unittest.TestCase ): """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=7 , lowerCAmelCase__=3 , lowerCAmelCase__=3_0 , lowerCAmelCase__=4_0_0 , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=0.9 , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__=[0.5, 0.5, 0.5] , lowerCAmelCase__=[0.5, 0.5, 0.5] , ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = size if size is not None else {'shortest_edge': 3_0} SCREAMING_SNAKE_CASE_ : Tuple = crop_size if crop_size is not None else {'height': 3_0, 'width': 3_0} SCREAMING_SNAKE_CASE_ : List[str] = parent SCREAMING_SNAKE_CASE_ : str = batch_size SCREAMING_SNAKE_CASE_ : List[Any] = num_channels SCREAMING_SNAKE_CASE_ : Union[str, Any] = min_resolution SCREAMING_SNAKE_CASE_ : Union[str, Any] = max_resolution SCREAMING_SNAKE_CASE_ : Optional[Any] = do_resize_and_center_crop SCREAMING_SNAKE_CASE_ : Any = size SCREAMING_SNAKE_CASE_ : int = crop_pct SCREAMING_SNAKE_CASE_ : Any = crop_size SCREAMING_SNAKE_CASE_ : str = do_normalize SCREAMING_SNAKE_CASE_ : Dict = image_mean SCREAMING_SNAKE_CASE_ : Any = image_std def UpperCamelCase__ ( self ): """simple docstring""" return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class __lowercase (__UpperCamelCase , unittest.TestCase ): """simple docstring""" _UpperCAmelCase = PoolFormerImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = PoolFormerImageProcessingTester(self ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_resize_and_center_crop' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'size' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'crop_pct' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'do_normalize' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'image_mean' ) ) self.assertTrue(hasattr(lowerCAmelCase__ , 'image_std' ) ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 3_0} ) self.assertEqual(image_processor.crop_size , {'height': 3_0, 'width': 3_0} ) SCREAMING_SNAKE_CASE_ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 , crop_size=8_4 ) self.assertEqual(image_processor.size , {'shortest_edge': 4_2} ) self.assertEqual(image_processor.crop_size , {'height': 8_4, 'width': 8_4} ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE_ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE_ : Tuple = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE_ : List[Any] = image_processing(lowerCAmelCase__ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE_ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , numpify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE_ : Union[str, Any] = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE_ : Optional[Any] = image_processing(lowerCAmelCase__ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE_ : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase__ , torchify=lowerCAmelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase__ , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE_ : int = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched SCREAMING_SNAKE_CASE_ : Optional[int] = image_processing(lowerCAmelCase__ , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
351
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ : Optional[Any] =logging.get_logger(__name__) lowerCAmelCase__ : Optional[int] ={ 'sail/poolformer_s12': 'https://huggingface.co/sail/poolformer_s12/resolve/main/config.json', # See all PoolFormer models at https://huggingface.co/models?filter=poolformer } class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase = """poolformer""" def __init__( self , lowerCAmelCase__=3 , lowerCAmelCase__=1_6 , lowerCAmelCase__=1_6 , lowerCAmelCase__=3 , lowerCAmelCase__=4.0 , lowerCAmelCase__=[2, 2, 6, 2] , lowerCAmelCase__=[6_4, 1_2_8, 3_2_0, 5_1_2] , lowerCAmelCase__=[7, 3, 3, 3] , lowerCAmelCase__=[4, 2, 2, 2] , lowerCAmelCase__=[2, 1, 1, 1] , lowerCAmelCase__=4 , lowerCAmelCase__=0.0 , lowerCAmelCase__="gelu" , lowerCAmelCase__=True , lowerCAmelCase__=1E-5 , lowerCAmelCase__=0.02 , **lowerCAmelCase__ , ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = num_channels SCREAMING_SNAKE_CASE_ : List[str] = patch_size SCREAMING_SNAKE_CASE_ : Tuple = stride SCREAMING_SNAKE_CASE_ : List[Any] = padding SCREAMING_SNAKE_CASE_ : Union[str, Any] = pool_size SCREAMING_SNAKE_CASE_ : Tuple = hidden_sizes SCREAMING_SNAKE_CASE_ : Tuple = mlp_ratio SCREAMING_SNAKE_CASE_ : Dict = depths SCREAMING_SNAKE_CASE_ : List[Any] = patch_sizes SCREAMING_SNAKE_CASE_ : List[Any] = strides SCREAMING_SNAKE_CASE_ : int = num_encoder_blocks SCREAMING_SNAKE_CASE_ : List[Any] = drop_path_rate SCREAMING_SNAKE_CASE_ : Tuple = hidden_act SCREAMING_SNAKE_CASE_ : str = use_layer_scale SCREAMING_SNAKE_CASE_ : List[str] = layer_scale_init_value SCREAMING_SNAKE_CASE_ : Tuple = initializer_range super().__init__(**lowerCAmelCase__ ) class __lowercase (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCAmelCase = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 2E-3
162
0
import argparse import os import re import packaging.version _snake_case = '''examples/''' _snake_case = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } _snake_case = { '''init''': '''src/diffusers/__init__.py''', '''setup''': '''setup.py''', } _snake_case = '''README.md''' def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE_ , "r" , encoding="utf-8" , newline="\n" ) as f: lowerCamelCase : int = f.read() lowerCamelCase , lowerCamelCase : Union[str, Any] = REPLACE_PATTERNS[pattern] lowerCamelCase : Union[str, Any] = replace.replace("VERSION" , SCREAMING_SNAKE_CASE_ ) lowerCamelCase : str = re_pattern.sub(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" , newline="\n" ) as f: f.write(SCREAMING_SNAKE_CASE_ ) def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' for folder, directories, fnames in os.walk(SCREAMING_SNAKE_CASE_ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("research_projects" ) if "legacy" in directories: directories.remove("legacy" ) for fname in fnames: if fname.endswith(".py" ): update_version_in_file(os.path.join(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ , pattern="examples" ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if not patch: update_version_in_examples(SCREAMING_SNAKE_CASE_ ) def lowercase_( ): '''simple docstring''' lowerCamelCase : Union[str, Any] = "🤗 Transformers currently provides the following architectures" lowerCamelCase : List[Any] = "1. Want to contribute a new model?" with open(SCREAMING_SNAKE_CASE_ , "r" , encoding="utf-8" , newline="\n" ) as f: lowerCamelCase : Any = f.readlines() # Find the start of the list. lowerCamelCase : Union[str, Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowerCamelCase : List[str] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("1." ): lowerCamelCase : Any = lines[index].replace( "https://huggingface.co/docs/diffusers/main/model_doc" , "https://huggingface.co/docs/diffusers/model_doc" , ) index += 1 with open(SCREAMING_SNAKE_CASE_ , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(SCREAMING_SNAKE_CASE_ ) def lowercase_( ): '''simple docstring''' with open(REPLACE_FILES["init"] , "r" ) as f: lowerCamelCase : Optional[Any] = f.read() lowerCamelCase : Optional[int] = REPLACE_PATTERNS["init"][0].search(SCREAMING_SNAKE_CASE_ ).groups()[0] return packaging.version.parse(SCREAMING_SNAKE_CASE_ ) def lowercase_( SCREAMING_SNAKE_CASE_=False ): '''simple docstring''' lowerCamelCase : Dict = get_version() if patch and default_version.is_devrelease: raise ValueError("Can't create a patch version from the dev branch, checkout a released version!" ) if default_version.is_devrelease: lowerCamelCase : Dict = default_version.base_version elif patch: lowerCamelCase : List[Any] = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: lowerCamelCase : List[str] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. lowerCamelCase : Any = input(f"""Which version are you releasing? [{default_version}]""" ) if len(SCREAMING_SNAKE_CASE_ ) == 0: lowerCamelCase : Union[str, Any] = default_version print(f"""Updating version to {version}.""" ) global_version_update(SCREAMING_SNAKE_CASE_ , patch=SCREAMING_SNAKE_CASE_ ) def lowercase_( ): '''simple docstring''' lowerCamelCase : Optional[Any] = get_version() lowerCamelCase : Union[str, Any] = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" lowerCamelCase : List[Any] = current_version.base_version # Check with the user we got that right. lowerCamelCase : int = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(SCREAMING_SNAKE_CASE_ ) == 0: lowerCamelCase : Union[str, Any] = dev_version print(f"""Updating version to {version}.""" ) global_version_update(SCREAMING_SNAKE_CASE_ ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') _snake_case = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
283
import argparse import torch from transformers import YosoConfig, YosoForMaskedLM def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' if "model" in orig_key: lowerCamelCase : Dict = orig_key.replace("model." , "" ) if "norm1" in orig_key: lowerCamelCase : Union[str, Any] = orig_key.replace("norm1" , "attention.output.LayerNorm" ) if "norm2" in orig_key: lowerCamelCase : Union[str, Any] = orig_key.replace("norm2" , "output.LayerNorm" ) if "norm" in orig_key: lowerCamelCase : Optional[Any] = orig_key.replace("norm" , "LayerNorm" ) if "transformer" in orig_key: lowerCamelCase : int = orig_key.split("." )[0].split("_" )[-1] lowerCamelCase : Dict = orig_key.replace(f"""transformer_{layer_num}""" , f"""encoder.layer.{layer_num}""" ) if "mha.attn" in orig_key: lowerCamelCase : List[str] = orig_key.replace("mha.attn" , "attention.self" ) if "mha" in orig_key: lowerCamelCase : List[Any] = orig_key.replace("mha" , "attention" ) if "W_q" in orig_key: lowerCamelCase : Optional[int] = orig_key.replace("W_q" , "self.query" ) if "W_k" in orig_key: lowerCamelCase : List[Any] = orig_key.replace("W_k" , "self.key" ) if "W_v" in orig_key: lowerCamelCase : Union[str, Any] = orig_key.replace("W_v" , "self.value" ) if "ff1" in orig_key: lowerCamelCase : Union[str, Any] = orig_key.replace("ff1" , "intermediate.dense" ) if "ff2" in orig_key: lowerCamelCase : Optional[int] = orig_key.replace("ff2" , "output.dense" ) if "ff" in orig_key: lowerCamelCase : Optional[int] = orig_key.replace("ff" , "output.dense" ) if "mlm_class" in orig_key: lowerCamelCase : Dict = orig_key.replace("mlm.mlm_class" , "cls.predictions.decoder" ) if "mlm" in orig_key: lowerCamelCase : List[Any] = orig_key.replace("mlm" , "cls.predictions.transform" ) if "cls" not in orig_key: lowerCamelCase : int = "yoso." + orig_key return orig_key def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' for key in orig_state_dict.copy().keys(): lowerCamelCase : List[str] = orig_state_dict.pop(SCREAMING_SNAKE_CASE_ ) if ("pooler" in key) or ("sen_class" in key): continue else: lowerCamelCase : Dict = val lowerCamelCase : Dict = orig_state_dict["cls.predictions.decoder.bias"] lowerCamelCase : Dict = torch.arange(SCREAMING_SNAKE_CASE_ ).expand((1, -1) ) + 2 return orig_state_dict def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : List[Any] = torch.load(SCREAMING_SNAKE_CASE_ , map_location="cpu" )["model_state_dict"] lowerCamelCase : List[str] = YosoConfig.from_json_file(SCREAMING_SNAKE_CASE_ ) lowerCamelCase : Any = YosoForMaskedLM(SCREAMING_SNAKE_CASE_ ) lowerCamelCase : List[Any] = convert_checkpoint_helper(config.max_position_embeddings , SCREAMING_SNAKE_CASE_ ) print(model.load_state_dict(SCREAMING_SNAKE_CASE_ ) ) model.eval() model.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(f"""Checkpoint successfuly converted. Model saved at {pytorch_dump_path}""" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--pytorch_model_path''', default=None, type=str, required=True, help='''Path to YOSO pytorch checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The json file for YOSO model config.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) _snake_case = parser.parse_args() convert_yoso_checkpoint(args.pytorch_model_path, args.config_file, args.pytorch_dump_path)
283
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a_ : Dict = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Dict = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys a_ : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
368
'''simple docstring''' def _A (lowerCAmelCase__ :list ) -> float: '''simple docstring''' _a = 0 while len(lowerCAmelCase__ ) > 1: _a = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): _a = files.index(min(lowerCAmelCase__ ) ) temp += files[min_index] files.pop(lowerCAmelCase__ ) files.append(lowerCAmelCase__ ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
104
0
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor lowercase : Optional[Any] = logging.get_logger(__name__) class lowerCamelCase__ ( _lowerCAmelCase): '''simple docstring''' def __init__( self :Optional[Any] , *a :List[str] , **a :int ) -> int: warnings.warn( "The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DPTImageProcessor instead." , _lowercase , ) super().__init__(*_lowercase , **_lowercase )
232
"""simple docstring""" from __future__ import annotations def lowercase__ ( snake_case_ :list[float] , snake_case_ :list[float] ): __UpperCAmelCase = sorted(numsa + numsa ) __UpperCAmelCase , __UpperCAmelCase = divmod(len(snake_case_ ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() _lowercase : int = [float(x) for x in input('Enter the elements of first array: ').split()] _lowercase : Tuple = [float(x) for x in input('Enter the elements of second array: ').split()] print(f"""The median of two arrays is: {median_of_two_arrays(array_a, array_a)}""")
332
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case : Dict = { """configuration_lxmert""": ["""LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LxmertConfig"""], """tokenization_lxmert""": ["""LxmertTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Union[str, Any] = ["""LxmertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Tuple = [ """LxmertEncoder""", """LxmertForPreTraining""", """LxmertForQuestionAnswering""", """LxmertModel""", """LxmertPreTrainedModel""", """LxmertVisualFeatureEncoder""", """LxmertXLayer""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : int = [ """TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFLxmertForPreTraining""", """TFLxmertMainLayer""", """TFLxmertModel""", """TFLxmertPreTrainedModel""", """TFLxmertVisualFeatureEncoder""", ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys __snake_case : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
359
import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json __snake_case : Tuple = """sshleifer/mar_enro_6_3_student""" class __SCREAMING_SNAKE_CASE ( __lowercase): def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() lowerCAmelCase__ = cached_path( 'https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz' , extract_compressed_file=_UpperCamelCase , ) lowerCAmelCase__ = F"{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k" @slow @require_torch_gpu def UpperCamelCase__ ( self ): """simple docstring""" MarianMTModel.from_pretrained(_UpperCamelCase ) @slow @require_torch_gpu def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = { '$MAX_LEN': 64, '$BS': 64, '$GAS': 1, '$ENRO_DIR': self.data_dir, 'facebook/mbart-large-cc25': MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", '--learning_rate=3e-5': '--learning_rate 3e-4', '--num_train_epochs 6': '--num_train_epochs 1', } # Clean up bash script lowerCAmelCase__ = (self.test_file_dir / 'train_mbart_cc25_enro.sh').open().read().split('finetune.py' )[1].strip() lowerCAmelCase__ = bash_script.replace('\\\n' , '' ).strip().replace('"$@"' , '' ) for k, v in env_vars_to_replace.items(): lowerCAmelCase__ = bash_script.replace(_UpperCamelCase , str(_UpperCamelCase ) ) lowerCAmelCase__ = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") lowerCAmelCase__ = F"\n --output_dir {output_dir}\n --tokenizer_name Helsinki-NLP/opus-mt-en-ro\n --sortish_sampler\n --do_predict\n --gpus 1\n --freeze_encoder\n --n_train 40000\n --n_val 500\n --n_test 500\n --fp16_opt_level O1\n --num_sanity_val_steps 0\n --eval_beams 2\n ".split() # XXX: args.gpus > 1 : handle multi_gpu in the future lowerCAmelCase__ = ['finetune.py'] + bash_script.split() + args with patch.object(_UpperCamelCase , 'argv' , _UpperCamelCase ): lowerCAmelCase__ = argparse.ArgumentParser() lowerCAmelCase__ = pl.Trainer.add_argparse_args(_UpperCamelCase ) lowerCAmelCase__ = SummarizationModule.add_model_specific_args(_UpperCamelCase , os.getcwd() ) lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = main(_UpperCamelCase ) # Check metrics lowerCAmelCase__ = load_json(model.metrics_save_path ) lowerCAmelCase__ = metrics['val'][0] lowerCAmelCase__ = metrics['val'][-1] self.assertEqual(len(metrics['val'] ) , (args.max_epochs / args.val_check_interval) ) assert isinstance(last_step_stats[F"val_avg_{model.val_metric}"] , _UpperCamelCase ) self.assertGreater(last_step_stats['val_avg_gen_time'] , 0.01 ) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats['val_avg_gen_time'] , 1.0 ) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats['val_avg_bleu'] - first_step_stats['val_avg_bleu'] , 2 ) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats['val_avg_bleu'] , 17 ) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics['val'][-1]['val_avg_bleu'] - metrics['test'][-1]['test_avg_bleu'] ) , 1.1 ) # check lightning ckpt can be loaded and has a reasonable statedict lowerCAmelCase__ = os.listdir(_UpperCamelCase ) lowerCAmelCase__ = [x for x in contents if x.endswith('.ckpt' )][0] lowerCAmelCase__ = os.path.join(args.output_dir , _UpperCamelCase ) lowerCAmelCase__ = torch.load(_UpperCamelCase , map_location='cpu' ) lowerCAmelCase__ = 'model.model.decoder.layers.0.encoder_attn_layer_norm.weight' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowerCAmelCase__ = {os.path.basename(_UpperCamelCase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['test'] ) == 1 class __SCREAMING_SNAKE_CASE ( __lowercase): @timeout_decorator.timeout(6_00 ) @slow @require_torch_gpu def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = F"{self.test_file_dir_str}/test_data/wmt_en_ro" lowerCAmelCase__ = { '--fp16_opt_level=O1': '', '$MAX_LEN': 1_28, '$BS': 16, '$GAS': 1, '$ENRO_DIR': data_dir, '$m': 'sshleifer/student_marian_en_ro_6_1', 'val_check_interval=0.25': 'val_check_interval=1.0', } # Clean up bash script lowerCAmelCase__ = ( (self.test_file_dir / 'distil_marian_no_teacher.sh').open().read().split('distillation.py' )[1].strip() ) lowerCAmelCase__ = bash_script.replace('\\\n' , '' ).strip().replace('"$@"' , '' ) lowerCAmelCase__ = bash_script.replace('--fp16 ' , ' ' ) for k, v in env_vars_to_replace.items(): lowerCAmelCase__ = bash_script.replace(_UpperCamelCase , str(_UpperCamelCase ) ) lowerCAmelCase__ = self.get_auto_remove_tmp_dir() lowerCAmelCase__ = bash_script.replace('--fp16' , '' ) lowerCAmelCase__ = 6 lowerCAmelCase__ = ( ['distillation.py'] + bash_script.split() + [ F"--output_dir={output_dir}", '--gpus=1', '--learning_rate=1e-3', F"--num_train_epochs={epochs}", '--warmup_steps=10', '--val_check_interval=1.0', '--do_predict', ] ) with patch.object(_UpperCamelCase , 'argv' , _UpperCamelCase ): lowerCAmelCase__ = argparse.ArgumentParser() lowerCAmelCase__ = pl.Trainer.add_argparse_args(_UpperCamelCase ) lowerCAmelCase__ = SummarizationDistiller.add_model_specific_args(_UpperCamelCase , os.getcwd() ) lowerCAmelCase__ = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu lowerCAmelCase__ = distill_main(_UpperCamelCase ) # Check metrics lowerCAmelCase__ = load_json(model.metrics_save_path ) lowerCAmelCase__ = metrics['val'][0] lowerCAmelCase__ = metrics['val'][-1] assert len(metrics['val'] ) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.01 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[F"val_avg_{model.val_metric}"] , _UpperCamelCase ) # check lightning ckpt can be loaded and has a reasonable statedict lowerCAmelCase__ = os.listdir(_UpperCamelCase ) lowerCAmelCase__ = [x for x in contents if x.endswith('.ckpt' )][0] lowerCAmelCase__ = os.path.join(args.output_dir , _UpperCamelCase ) lowerCAmelCase__ = torch.load(_UpperCamelCase , map_location='cpu' ) lowerCAmelCase__ = 'model.model.decoder.layers.0.encoder_attn_layer_norm.weight' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowerCAmelCase__ = {os.path.basename(_UpperCamelCase ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['test'] ) == 1
122
0
from __future__ import annotations def snake_case_ ( snake_case , snake_case ) -> list[str]: if nth_term == "": return [""] lowercase__: Tuple = int(snake_case ) lowercase__: int = int(snake_case ) lowercase__: list[str] = [] for temp in range(int(snake_case ) ): series.append(f'1 / {pow(temp + 1 , int(snake_case ) )}' if series else '1' ) return series if __name__ == "__main__": import doctest doctest.testmod() __lowerCAmelCase = int(input('''Enter the last number (nth term) of the P-Series''')) __lowerCAmelCase = int(input('''Enter the power for P-Series''')) print('''Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p''') print(p_series(nth_term, power))
196
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class __a : @staticmethod def SCREAMING_SNAKE_CASE__ ( *lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: '''simple docstring''' pass def snake_case_ ( snake_case ) -> Optional[Any]: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. __lowerCAmelCase = ( '''https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png''' ) @is_pipeline_test @require_torch @require_vision class __a ( unittest.TestCase ): __lowercase : Dict = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: '''simple docstring''' lowercase__: Optional[Any] = pipeline( 'document-question-answering' , model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ , image_processor=lowerCAmelCase__ ) lowercase__: int = INVOICE_URL lowercase__: Dict = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , '' ) ) ) lowercase__: str = 'What is the placebo?' lowercase__: Any = [ { 'image': load_image(lowerCAmelCase__ ), 'question': question, }, { 'image': image, 'question': question, }, { 'image': image, 'question': question, 'word_boxes': word_boxes, }, ] return dqa_pipeline, examples def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: '''simple docstring''' lowercase__: str = dqa_pipeline(lowerCAmelCase__ , top_k=2 ) self.assertEqual( lowerCAmelCase__ , [ [ {'score': ANY(lowerCAmelCase__ ), 'answer': ANY(lowerCAmelCase__ ), 'start': ANY(lowerCAmelCase__ ), 'end': ANY(lowerCAmelCase__ )}, {'score': ANY(lowerCAmelCase__ ), 'answer': ANY(lowerCAmelCase__ ), 'start': ANY(lowerCAmelCase__ ), 'end': ANY(lowerCAmelCase__ )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' lowercase__: Union[str, Any] = pipeline('document-question-answering' , model='hf-internal-testing/tiny-random-layoutlmv2' ) lowercase__: Optional[Any] = INVOICE_URL lowercase__: int = 'How many cats are there?' lowercase__: List[str] = [ {'score': 0.0_0_0_1, 'answer': 'oy 2312/2019', 'start': 38, 'end': 39}, {'score': 0.0_0_0_1, 'answer': 'oy 2312/2019 DUE', 'start': 38, 'end': 40}, ] lowercase__: Dict = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , lowerCAmelCase__ ) lowercase__: Tuple = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , lowerCAmelCase__ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably lowercase__: str = './tests/fixtures/tests_samples/COCO/000000039769.png' lowercase__: Tuple = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(lowerCAmelCase__ , [] ) # We can optionnally pass directly the words and bounding boxes lowercase__: int = './tests/fixtures/tests_samples/COCO/000000039769.png' lowercase__: List[Any] = [] lowercase__: Optional[int] = [] lowercase__: Any = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , words=lowerCAmelCase__ , boxes=lowerCAmelCase__ , top_k=2 ) self.assertEqual(lowerCAmelCase__ , [] ) @slow @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE__ ( self ) -> Any: '''simple docstring''' lowercase__: List[str] = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , ) lowercase__: int = INVOICE_URL lowercase__: str = 'What is the invoice number?' lowercase__: Union[str, Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_4_4, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_0_0_9, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Any = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_4_4, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_0_0_9, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Optional[int] = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.9_9_4_4, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_0_0_9, 'answer': 'us-001', 'start': 16, 'end': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' lowercase__: Any = pipeline( 'document-question-answering' , model='tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa' , revision='9977165' , max_seq_len=50 , ) lowercase__: Optional[int] = INVOICE_URL lowercase__: Union[str, Any] = 'What is the invoice number?' lowercase__: Optional[Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_7_4, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9_9_4_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Tuple = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_7_4, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9_9_4_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Dict = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.9_9_7_4, 'answer': '1110212019', 'start': 23, 'end': 23}, {'score': 0.9_9_4_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' lowercase__: Optional[Any] = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCAmelCase__ ) lowercase__: Optional[Any] = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCAmelCase__ , revision='3dc6de3' , ) lowercase__: List[str] = INVOICE_URL lowercase__: Union[str, Any] = 'What is the invoice number?' lowercase__: Dict = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowercase__: List[str] = dqa_pipeline({'image': image, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) lowercase__: int = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] ] * 2 , ) lowercase__: Any = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , '' ) ) ) # This model should also work if `image` is set to None lowercase__: List[Any] = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.4_2_5_1, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.0_8_1_9, 'answer': '1110212019', 'start': 23, 'end': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' lowercase__: Any = AutoTokenizer.from_pretrained( 'impira/layoutlm-document-qa' , revision='3dc6de3' , add_prefix_space=lowerCAmelCase__ ) lowercase__: str = pipeline( 'document-question-answering' , model='impira/layoutlm-document-qa' , tokenizer=lowerCAmelCase__ , revision='3dc6de3' , max_seq_len=50 , ) lowercase__: Optional[Any] = INVOICE_URL lowercase__: Optional[Any] = 'What is the invoice number?' lowercase__: Optional[int] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_9_9, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9_9_9_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) lowercase__: Any = dqa_pipeline( [{'image': image, 'question': question}, {'image': image, 'question': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {'score': 0.9_9_9_9, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9_9_9_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] ] * 2 , ) lowercase__: Optional[int] = list(zip(*apply_tesseract(load_image(lowerCAmelCase__ ) , lowerCAmelCase__ , '' ) ) ) # This model should also work if `image` is set to None lowercase__: Tuple = dqa_pipeline({'image': None, 'word_boxes': word_boxes, 'question': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {'score': 0.9_9_9_9, 'answer': 'us-001', 'start': 16, 'end': 16}, {'score': 0.9_9_9_8, 'answer': 'us-001', 'start': 16, 'end': 16}, ] , ) @slow @require_torch def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[Any]: '''simple docstring''' lowercase__: List[Any] = pipeline( 'document-question-answering' , model='naver-clova-ix/donut-base-finetuned-docvqa' , tokenizer=AutoTokenizer.from_pretrained('naver-clova-ix/donut-base-finetuned-docvqa' ) , feature_extractor='naver-clova-ix/donut-base-finetuned-docvqa' , ) lowercase__: int = INVOICE_URL lowercase__: int = 'What is the invoice number?' lowercase__: Union[str, Any] = dqa_pipeline(image=lowerCAmelCase__ , question=lowerCAmelCase__ , top_k=2 ) self.assertEqual(nested_simplify(lowerCAmelCase__ , decimals=4 ) , [{'answer': 'us-001'}] ) @require_tf @unittest.skip('Document question answering not implemented in TF' ) def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' pass
196
1
'''simple docstring''' import sys __snake_case : Union[str, Any] = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def _UpperCAmelCase ( _UpperCamelCase : str ) -> int: A_ = 1 for digit in s: product *= int(_UpperCamelCase ) return product def _UpperCAmelCase ( _UpperCamelCase : str = N ) -> int: A_ = -sys.maxsize - 1 A_ = n[:13] A_ = 13 while cur_index < len(_UpperCamelCase ) - 13: if int(n[cur_index] ) >= int(substr[0] ): A_ = substr[1:] + n[cur_index] cur_index += 1 else: A_ = max(_UpperCamelCase, str_eval(_UpperCamelCase ) ) A_ = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(F"""{solution() = }""")
18
'''simple docstring''' from statistics import mean, stdev def _UpperCAmelCase ( _UpperCamelCase : list, _UpperCamelCase : int = 3 ) -> list: A_ = min(_UpperCamelCase ) A_ = max(_UpperCamelCase ) # normalize data return [round((x - x_min) / (x_max - x_min), _UpperCamelCase ) for x in data] def _UpperCAmelCase ( _UpperCamelCase : list, _UpperCamelCase : int = 3 ) -> list: A_ = mean(_UpperCamelCase ) A_ = stdev(_UpperCamelCase ) # standardize data return [round((x - mu) / (sigma), _UpperCamelCase ) for x in data]
18
1
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class _UpperCAmelCase ( lowercase__ ): """simple docstring""" lowercase__ = (CMStochasticIterativeScheduler,) lowercase__ = 10 def lowercase__ ( self : List[Any], **lowerCamelCase : int ): '''simple docstring''' lowercase__ = { "num_train_timesteps": 201, "sigma_min": 0.002, "sigma_max": 80.0, } config.update(**lowerCamelCase ) return config def lowercase__ ( self : str ): '''simple docstring''' lowercase__ = 10 lowercase__ = self.get_scheduler_config() lowercase__ = self.scheduler_classes[0](**lowerCamelCase ) scheduler.set_timesteps(lowerCamelCase ) lowercase__ = scheduler.timesteps[0] lowercase__ = scheduler.timesteps[1] lowercase__ = self.dummy_sample lowercase__ = 0.1 * sample lowercase__ = scheduler.step(lowerCamelCase, lowerCamelCase, lowerCamelCase ).prev_sample lowercase__ = scheduler.step(lowerCamelCase, lowerCamelCase, lowerCamelCase ).prev_sample self.assertEqual(output_a.shape, sample.shape ) self.assertEqual(output_a.shape, output_a.shape ) def lowercase__ ( self : Dict ): '''simple docstring''' for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=lowerCamelCase ) def lowercase__ ( self : List[str] ): '''simple docstring''' for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=lowerCamelCase ) def lowercase__ ( self : Union[str, Any] ): '''simple docstring''' lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**lowerCamelCase ) lowercase__ = 1 scheduler.set_timesteps(lowerCamelCase ) lowercase__ = scheduler.timesteps lowercase__ = torch.manual_seed(0 ) lowercase__ = self.dummy_model() lowercase__ = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(lowerCamelCase ): # 1. scale model input lowercase__ = scheduler.scale_model_input(lowerCamelCase, lowerCamelCase ) # 2. predict noise residual lowercase__ = model(lowerCamelCase, lowerCamelCase ) # 3. predict previous sample x_t-1 lowercase__ = scheduler.step(lowerCamelCase, lowerCamelCase, lowerCamelCase, generator=lowerCamelCase ).prev_sample lowercase__ = pred_prev_sample lowercase__ = torch.sum(torch.abs(lowerCamelCase ) ) lowercase__ = torch.mean(torch.abs(lowerCamelCase ) ) assert abs(result_sum.item() - 192.7614 ) < 1E-2 assert abs(result_mean.item() - 0.2510 ) < 1E-3 def lowercase__ ( self : Optional[Any] ): '''simple docstring''' lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**lowerCamelCase ) lowercase__ = [106, 0] scheduler.set_timesteps(timesteps=lowerCamelCase ) lowercase__ = scheduler.timesteps lowercase__ = torch.manual_seed(0 ) lowercase__ = self.dummy_model() lowercase__ = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input lowercase__ = scheduler.scale_model_input(lowerCamelCase, lowerCamelCase ) # 2. predict noise residual lowercase__ = model(lowerCamelCase, lowerCamelCase ) # 3. predict previous sample x_t-1 lowercase__ = scheduler.step(lowerCamelCase, lowerCamelCase, lowerCamelCase, generator=lowerCamelCase ).prev_sample lowercase__ = pred_prev_sample lowercase__ = torch.sum(torch.abs(lowerCamelCase ) ) lowercase__ = torch.mean(torch.abs(lowerCamelCase ) ) assert abs(result_sum.item() - 347.6357 ) < 1E-2 assert abs(result_mean.item() - 0.4527 ) < 1E-3 def lowercase__ ( self : Optional[int] ): '''simple docstring''' lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**lowerCamelCase ) lowercase__ = [39, 30, 12, 15, 0] with self.assertRaises(lowerCamelCase, msg='''`timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=lowerCamelCase ) def lowercase__ ( self : Union[str, Any] ): '''simple docstring''' lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**lowerCamelCase ) lowercase__ = [39, 30, 12, 1, 0] lowercase__ = len(lowerCamelCase ) with self.assertRaises(lowerCamelCase, msg='''Can only pass one of `num_inference_steps` or `timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=lowerCamelCase, timesteps=lowerCamelCase ) def lowercase__ ( self : Union[str, Any] ): '''simple docstring''' lowercase__ = self.scheduler_classes[0] lowercase__ = self.get_scheduler_config() lowercase__ = scheduler_class(**lowerCamelCase ) lowercase__ = [scheduler.config.num_train_timesteps] with self.assertRaises( lowerCamelCase, msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''', ): scheduler.set_timesteps(timesteps=lowerCamelCase )
207
'''simple docstring''' a : List[str] = "0.21.0" from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
311
0
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class _snake_case ( unittest.TestCase ): def __init__( self: List[Any] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: List[str]=7 , __lowerCamelCase: Union[str, Any]=3 , __lowerCamelCase: Optional[Any]=18 , __lowerCamelCase: Tuple=30 , __lowerCamelCase: int=4_00 , __lowerCamelCase: Optional[Any]=True , __lowerCamelCase: Any=None , __lowerCamelCase: str=True , __lowerCamelCase: Any=[0.5, 0.5, 0.5] , __lowerCamelCase: Optional[int]=[0.5, 0.5, 0.5] , ) -> Dict: __UpperCAmelCase : Optional[int] = size if size is not None else {"height": 18, "width": 18} __UpperCAmelCase : List[Any] = parent __UpperCAmelCase : Any = batch_size __UpperCAmelCase : Union[str, Any] = num_channels __UpperCAmelCase : Any = image_size __UpperCAmelCase : Optional[int] = min_resolution __UpperCAmelCase : Optional[int] = max_resolution __UpperCAmelCase : Tuple = do_resize __UpperCAmelCase : Optional[Any] = size __UpperCAmelCase : Union[str, Any] = do_normalize __UpperCAmelCase : Tuple = image_mean __UpperCAmelCase : Tuple = image_std def _lowerCamelCase ( self: List[Any] ) -> str: return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _snake_case ( _lowercase , unittest.TestCase ): lowerCamelCase__: Tuple = DPTImageProcessor if is_vision_available() else None def _lowerCamelCase ( self: Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase : Tuple = DPTImageProcessingTester(self ) @property def _lowerCamelCase ( self: Optional[Any] ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def _lowerCamelCase ( self: Optional[int] ) -> Optional[Any]: __UpperCAmelCase : Optional[int] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(__lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(__lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(__lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(__lowerCamelCase , "size" ) ) def _lowerCamelCase ( self: Tuple ) -> List[str]: __UpperCAmelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 18} ) __UpperCAmelCase : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) def _lowerCamelCase ( self: List[str] ) -> Dict: # Initialize image_processing __UpperCAmelCase : int = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __UpperCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , Image.Image ) # Test not batched input __UpperCAmelCase : Optional[int] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Tuple = image_processing(__lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def _lowerCamelCase ( self: Optional[Any] ) -> Any: # Initialize image_processing __UpperCAmelCase : Any = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __UpperCAmelCase : Dict = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , numpify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , np.ndarray ) # Test not batched input __UpperCAmelCase : Dict = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Any = image_processing(__lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def _lowerCamelCase ( self: Dict ) -> Tuple: # Initialize image_processing __UpperCAmelCase : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __UpperCAmelCase : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCamelCase , torchify=__lowerCamelCase ) for image in image_inputs: self.assertIsInstance(__lowerCamelCase , torch.Tensor ) # Test not batched input __UpperCAmelCase : Dict = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched __UpperCAmelCase : Optional[Any] = image_processing(__lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
342
import argparse import struct import unittest class _snake_case : def __init__( self: Tuple , __lowerCamelCase: bytes ) -> None: __UpperCAmelCase : Tuple = data # Initialize hash values __UpperCAmelCase : Any = [ 0x6_A_0_9_E_6_6_7, 0xB_B_6_7_A_E_8_5, 0x3_C_6_E_F_3_7_2, 0xA_5_4_F_F_5_3_A, 0x5_1_0_E_5_2_7_F, 0x9_B_0_5_6_8_8_C, 0x1_F_8_3_D_9_A_B, 0x5_B_E_0_C_D_1_9, ] # Initialize round constants __UpperCAmelCase : Dict = [ 0x4_2_8_A_2_F_9_8, 0x7_1_3_7_4_4_9_1, 0xB_5_C_0_F_B_C_F, 0xE_9_B_5_D_B_A_5, 0x3_9_5_6_C_2_5_B, 0x5_9_F_1_1_1_F_1, 0x9_2_3_F_8_2_A_4, 0xA_B_1_C_5_E_D_5, 0xD_8_0_7_A_A_9_8, 0x1_2_8_3_5_B_0_1, 0x2_4_3_1_8_5_B_E, 0x5_5_0_C_7_D_C_3, 0x7_2_B_E_5_D_7_4, 0x8_0_D_E_B_1_F_E, 0x9_B_D_C_0_6_A_7, 0xC_1_9_B_F_1_7_4, 0xE_4_9_B_6_9_C_1, 0xE_F_B_E_4_7_8_6, 0x0_F_C_1_9_D_C_6, 0x2_4_0_C_A_1_C_C, 0x2_D_E_9_2_C_6_F, 0x4_A_7_4_8_4_A_A, 0x5_C_B_0_A_9_D_C, 0x7_6_F_9_8_8_D_A, 0x9_8_3_E_5_1_5_2, 0xA_8_3_1_C_6_6_D, 0xB_0_0_3_2_7_C_8, 0xB_F_5_9_7_F_C_7, 0xC_6_E_0_0_B_F_3, 0xD_5_A_7_9_1_4_7, 0x0_6_C_A_6_3_5_1, 0x1_4_2_9_2_9_6_7, 0x2_7_B_7_0_A_8_5, 0x2_E_1_B_2_1_3_8, 0x4_D_2_C_6_D_F_C, 0x5_3_3_8_0_D_1_3, 0x6_5_0_A_7_3_5_4, 0x7_6_6_A_0_A_B_B, 0x8_1_C_2_C_9_2_E, 0x9_2_7_2_2_C_8_5, 0xA_2_B_F_E_8_A_1, 0xA_8_1_A_6_6_4_B, 0xC_2_4_B_8_B_7_0, 0xC_7_6_C_5_1_A_3, 0xD_1_9_2_E_8_1_9, 0xD_6_9_9_0_6_2_4, 0xF_4_0_E_3_5_8_5, 0x1_0_6_A_A_0_7_0, 0x1_9_A_4_C_1_1_6, 0x1_E_3_7_6_C_0_8, 0x2_7_4_8_7_7_4_C, 0x3_4_B_0_B_C_B_5, 0x3_9_1_C_0_C_B_3, 0x4_E_D_8_A_A_4_A, 0x5_B_9_C_C_A_4_F, 0x6_8_2_E_6_F_F_3, 0x7_4_8_F_8_2_E_E, 0x7_8_A_5_6_3_6_F, 0x8_4_C_8_7_8_1_4, 0x8_C_C_7_0_2_0_8, 0x9_0_B_E_F_F_F_A, 0xA_4_5_0_6_C_E_B, 0xB_E_F_9_A_3_F_7, 0xC_6_7_1_7_8_F_2, ] __UpperCAmelCase : List[Any] = self.preprocessing(self.data ) self.final_hash() @staticmethod def _lowerCamelCase ( __lowerCamelCase: bytes ) -> bytes: __UpperCAmelCase : List[str] = B"\x80" + (B"\x00" * (63 - (len(__lowerCamelCase ) + 8) % 64)) __UpperCAmelCase : int = struct.pack(">Q" , (len(__lowerCamelCase ) * 8) ) return data + padding + big_endian_integer def _lowerCamelCase ( self: Dict ) -> None: # Convert into blocks of 64 bytes __UpperCAmelCase : Dict = [ self.preprocessed_data[x : x + 64] for x in range(0 , len(self.preprocessed_data ) , 64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers __UpperCAmelCase : List[str] = list(struct.unpack(">16L" , __lowerCamelCase ) ) # add 48 0-ed integers words += [0] * 48 __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : Tuple = self.hashes for index in range(0 , 64 ): if index > 15: # modify the zero-ed indexes at the end of the array __UpperCAmelCase : Union[str, Any] = ( self.ror(words[index - 15] , 7 ) ^ self.ror(words[index - 15] , 18 ) ^ (words[index - 15] >> 3) ) __UpperCAmelCase : str = ( self.ror(words[index - 2] , 17 ) ^ self.ror(words[index - 2] , 19 ) ^ (words[index - 2] >> 10) ) __UpperCAmelCase : Union[str, Any] = ( words[index - 16] + sa + words[index - 7] + sa ) % 0x1_0_0_0_0_0_0_0_0 # Compression __UpperCAmelCase : Union[str, Any] = self.ror(__lowerCamelCase , 6 ) ^ self.ror(__lowerCamelCase , 11 ) ^ self.ror(__lowerCamelCase , 25 ) __UpperCAmelCase : Tuple = (e & f) ^ ((~e & 0xF_F_F_F_F_F_F_F) & g) __UpperCAmelCase : int = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0x1_0_0_0_0_0_0_0_0 __UpperCAmelCase : List[Any] = self.ror(__lowerCamelCase , 2 ) ^ self.ror(__lowerCamelCase , 13 ) ^ self.ror(__lowerCamelCase , 22 ) __UpperCAmelCase : Dict = (a & b) ^ (a & c) ^ (b & c) __UpperCAmelCase : int = (sa + maj) % 0x1_0_0_0_0_0_0_0_0 __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : int = ( g, f, e, ((d + tempa) % 0x1_0_0_0_0_0_0_0_0), c, b, a, ((tempa + tempa) % 0x1_0_0_0_0_0_0_0_0), ) __UpperCAmelCase : Optional[int] = [a, b, c, d, e, f, g, h] # Modify final values __UpperCAmelCase : List[str] = [ ((element + mutated_hash_values[index]) % 0x1_0_0_0_0_0_0_0_0) for index, element in enumerate(self.hashes ) ] __UpperCAmelCase : int = "".join([hex(__lowerCamelCase )[2:].zfill(8 ) for value in self.hashes] ) def _lowerCamelCase ( self: List[str] , __lowerCamelCase: int , __lowerCamelCase: int ) -> int: return 0xF_F_F_F_F_F_F_F & (value << (32 - rotations)) | (value >> rotations) class _snake_case ( unittest.TestCase ): def _lowerCamelCase ( self: List[Any] ) -> None: import hashlib __UpperCAmelCase : Dict = bytes("Test String" , "utf-8" ) self.assertEqual(SHAaaa(__lowerCamelCase ).hash , hashlib.shaaaa(__lowerCamelCase ).hexdigest() ) def _UpperCamelCase ( ) -> None: import doctest doctest.testmod() __UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( "-s", "--string", dest="input_string", default="Hello World!! Welcome to Cryptography", help="Hash the string", ) parser.add_argument( "-f", "--file", dest="input_file", help="Hash contents of a file" ) __UpperCAmelCase : List[Any] = parser.parse_args() __UpperCAmelCase : Optional[int] = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file, "rb" ) as f: __UpperCAmelCase : List[str] = f.read() else: __UpperCAmelCase : List[Any] = bytes(snake_case__, "utf-8" ) print(SHAaaa(snake_case__ ).hash ) if __name__ == "__main__": main()
342
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __a = {"configuration_speech_encoder_decoder": ["SpeechEncoderDecoderConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["SpeechEncoderDecoderModel"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["FlaxSpeechEncoderDecoderModel"] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
35
'''simple docstring''' import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings __a = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = field(default=_a , metadata={"help": "Whether to use SortishSampler or not."} ) lowercase = field( default=_a , metadata={"help": "Whether to use generate to calculate generative metrics (ROUGE, BLEU)."} ) lowercase = field( default=_a , metadata={ "help": ( "The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default " "to the `max_length` value of the model configuration." ) } , ) lowercase = field( default=_a , metadata={ "help": ( "The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default " "to the `num_beams` value of the model configuration." ) } , ) lowercase = field( default=_a , metadata={ "help": "Model id, file path or url pointing to a GenerationConfig json file, to use during prediction." } , ) def lowerCamelCase ( self : List[str] ): snake_case__ : int = super().to_dict() for k, v in d.items(): if isinstance(snake_case_ , snake_case_ ): snake_case__ : Optional[int] = v.to_dict() return d
35
1
import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor lowerCAmelCase__ :int = logging.getLogger(__name__) lowerCAmelCase__ :str = 5_0 # max width of layer names lowerCAmelCase__ :List[Any] = 7_0 # max width of quantizer names def lowerCAmelCase__ ( a__: Optional[int] ) -> int: '''simple docstring''' _UpperCAmelCase = parser.add_argument_group('quant_trainer arguments' ) group.add_argument('--wprec' , type=_A , default=8 , help='weight precision' ) group.add_argument('--aprec' , type=_A , default=8 , help='activation precision' ) group.add_argument('--quant-per-tensor' , action='store_true' , help='per tensor weight scaling' ) group.add_argument('--quant-disable' , action='store_true' , help='disable all quantizers' ) group.add_argument('--quant-disable-embeddings' , action='store_true' , help='disable all embeddings quantizers' ) group.add_argument('--quant-disable-keyword' , type=_A , nargs='+' , help='disable quantizers by keyword' ) group.add_argument('--quant-disable-layer-module' , type=_A , help='disable quantizers by keyword under layer.' ) group.add_argument('--quant-enable-layer-module' , type=_A , help='enable quantizers by keyword under layer' ) group.add_argument('--calibrator' , default='max' , help='which quantization range calibrator to use' ) group.add_argument('--percentile' , default=_A , type=_A , help='percentile for PercentileCalibrator' ) group.add_argument('--fuse-qkv' , action='store_true' , help='use the same scale factor for qkv' ) group.add_argument('--clip-gelu' , metavar='N' , type=_A , help='clip gelu output maximum value to N' ) group.add_argument( '--recalibrate-weights' , action='store_true' , help=( 'recalibrate weight amaxes by taking the max of the weights.' ' amaxes will be computed with the current quantization granularity (axis).' ) , ) def lowerCAmelCase__ ( a__: Tuple ) -> Tuple: '''simple docstring''' if args.calibrator == "max": _UpperCAmelCase = 'max' elif args.calibrator == "percentile": if args.percentile is None: raise ValueError('Specify --percentile when using percentile calibrator' ) _UpperCAmelCase = 'histogram' elif args.calibrator == "mse": _UpperCAmelCase = 'histogram' else: raise ValueError(F'''Invalid calibrator {args.calibrator}''' ) _UpperCAmelCase = QuantDescriptor(num_bits=args.aprec , calib_method=_A ) _UpperCAmelCase = QuantDescriptor(num_bits=args.wprec , axis=(None if args.quant_per_tensor else (0,)) ) quant_nn.QuantLinear.set_default_quant_desc_input(_A ) quant_nn.QuantLinear.set_default_quant_desc_weight(_A ) def lowerCAmelCase__ ( a__: Optional[int] , a__: Dict , a__: Dict=False , a__: Union[str, Any]=False ) -> Any: '''simple docstring''' logger.info('Configuring Model for Quantization' ) logger.info(F'''using quantization package {pytorch_quantization.__file__}''' ) if not calib: if args.quant_disable_embeddings: set_quantizer_by_name(_A , ['embeddings'] , which='weight' , _disabled=_A ) if args.quant_disable: set_quantizer_by_name(_A , [''] , _disabled=_A ) if args.quant_disable_keyword: set_quantizer_by_name(_A , args.quant_disable_keyword , _disabled=_A ) if args.quant_disable_layer_module: set_quantizer_by_name(_A , [R'layer.\d+.' + args.quant_disable_layer_module] , _disabled=_A ) if args.quant_enable_layer_module: set_quantizer_by_name(_A , [R'layer.\d+.' + args.quant_enable_layer_module] , _disabled=_A ) if args.recalibrate_weights: recalibrate_weights(_A ) if args.fuse_qkv: fuse_qkv(_A , _A ) if args.clip_gelu: clip_gelu(_A , args.clip_gelu ) # if args.local_rank in [-1, 0] and not calib: print_quant_summary(_A ) def lowerCAmelCase__ ( a__: Tuple ) -> Optional[int]: '''simple docstring''' logger.info('Enabling Calibration' ) for name, module in model.named_modules(): if name.endswith('_quantizer' ): if module._calibrator is not None: module.disable_quant() module.enable_calib() else: module.disable() logger.info(F'''{name:80}: {module}''' ) def lowerCAmelCase__ ( a__: List[str] , a__: Any ) -> Optional[int]: '''simple docstring''' logger.info('Loading calibrated amax' ) for name, module in model.named_modules(): if name.endswith('_quantizer' ): if module._calibrator is not None: if isinstance(module._calibrator , calib.MaxCalibrator ): module.load_calib_amax() else: module.load_calib_amax('percentile' , percentile=args.percentile ) module.enable_quant() module.disable_calib() else: module.enable() model.cuda() print_quant_summary(_A ) def lowerCAmelCase__ ( a__: Optional[int] , a__: int ) -> str: '''simple docstring''' def fusea(a__: Any , a__: Any , a__: str ): for mod in [qq, qk, qv]: if not hasattr(_A , '_amax' ): print(' WARNING: NO AMAX BUFFER' ) return _UpperCAmelCase = qq._amax.detach().item() _UpperCAmelCase = qk._amax.detach().item() _UpperCAmelCase = qv._amax.detach().item() _UpperCAmelCase = max(_A , _A , _A ) qq._amax.fill_(_A ) qk._amax.fill_(_A ) qv._amax.fill_(_A ) logger.info(F''' q={q:5.2f} k={k:5.2f} v={v:5.2f} -> {amax:5.2f}''' ) for name, mod in model.named_modules(): if name.endswith('.attention.self' ): logger.info(F'''FUSE_QKV: {name:{name_width}}''' ) fusea(mod.matmul_q_input_quantizer , mod.matmul_k_input_quantizer , mod.matmul_v_input_quantizer ) if args.quant_per_tensor: fusea(mod.query._weight_quantizer , mod.key._weight_quantizer , mod.value._weight_quantizer ) def lowerCAmelCase__ ( a__: Optional[Any] , a__: Dict ) -> int: '''simple docstring''' for name, mod in model.named_modules(): if name.endswith('.output.dense' ) and not name.endswith('attention.output.dense' ): _UpperCAmelCase = mod._input_quantizer._amax.data.detach().item() mod._input_quantizer._amax.data.detach().clamp_(max=_A ) _UpperCAmelCase = mod._input_quantizer._amax.data.detach().item() logger.info(F'''CLIP_GELU: {name:{name_width}} amax: {amax_init:5.2f} -> {amax:5.2f}''' ) def lowerCAmelCase__ ( a__: int ) -> str: '''simple docstring''' for name, mod in model.named_modules(): if hasattr(_A , '_weight_quantizer' ) and mod._weight_quantizer.axis is not None: _UpperCAmelCase = mod.weight.shape[0] _UpperCAmelCase = mod._weight_quantizer._amax.detach() _UpperCAmelCase = torch.ones(_A , dtype=amax.dtype , device=amax.device ) * amax print(F'''expanding {name} {amax} -> {mod._weight_quantizer._amax}''' ) def lowerCAmelCase__ ( a__: Dict ) -> Tuple: '''simple docstring''' for name, mod in model.named_modules(): if hasattr(_A , '_weight_quantizer' ): if not hasattr(mod.weight_quantizer , '_amax' ): print('RECALIB: {name:{name_width}} WARNING: NO AMAX BUFFER' ) continue # determine which axes to reduce across # e.g. a 4D tensor quantized per axis 0 should reduce over (1,2,3) _UpperCAmelCase = set() if mod._weight_quantizer.axis is None else set(mod._weight_quantizer.axis ) _UpperCAmelCase = set(range(len(mod.weight.size() ) ) ) - axis_set _UpperCAmelCase = pytorch_quantization.utils.reduce_amax(mod.weight , axis=_A , keepdims=_A ).detach() logger.info(F'''RECALIB: {name:{name_width}} {mod._weight_quantizer._amax.flatten()} -> {amax.flatten()}''' ) _UpperCAmelCase = amax def lowerCAmelCase__ ( a__: str , a__: List[str]=2_5 , a__: List[Any]=1_8_0 , a__: Union[str, Any]=None ) -> Dict: '''simple docstring''' if ignore is None: _UpperCAmelCase = [] elif not isinstance(_A , _A ): _UpperCAmelCase = [ignore] _UpperCAmelCase = 0 for name, mod in model.named_modules(): if not hasattr(_A , 'weight' ): continue _UpperCAmelCase = max(_A , len(_A ) ) for name, mod in model.named_modules(): _UpperCAmelCase = getattr(_A , '_input_quantizer' , _A ) _UpperCAmelCase = getattr(_A , '_weight_quantizer' , _A ) if not hasattr(_A , 'weight' ): continue if type(_A ) in ignore: continue if [True for s in ignore if type(_A ) is str and s in name]: continue _UpperCAmelCase = F'''Act:{input_q.extra_repr()}''' _UpperCAmelCase = F'''Wgt:{weight_q.extra_repr()}''' _UpperCAmelCase = F'''{name:{name_width}} {act_str} {wgt_str}''' if len(_A ) <= line_width: logger.info(_A ) else: logger.info(F'''{name:{name_width}} {act_str}''' ) logger.info(F'''{" ":{name_width}} {wgt_str}''' ) def lowerCAmelCase__ ( a__: List[Any] ) -> str: '''simple docstring''' _UpperCAmelCase = 0 for name, mod in model.named_modules(): if isinstance(_A , pytorch_quantization.nn.TensorQuantizer ): print(F'''{name:80} {mod}''' ) count += 1 print(F'''{count} TensorQuantizers found in model''' ) def lowerCAmelCase__ ( a__: Optional[Any] , a__: List[str] , a__: Dict , a__: Any , a__: str ) -> str: '''simple docstring''' _UpperCAmelCase = getattr(_A , _A , _A ) if quantizer_mod is not None: assert hasattr(_A , _A ) setattr(_A , _A , _A ) else: logger.warning(F'''{name} has no {quantizer}''' ) def lowerCAmelCase__ ( a__: Union[str, Any] , a__: Optional[Any] , a__: Optional[int]="both" , **a__: Optional[int] ) -> str: '''simple docstring''' _UpperCAmelCase = F'''Warning: changing {which} quantizers of {name:{qname_width}}''' for k, v in kwargs.items(): s += F''' {k}={v}''' if which in ["input", "both"]: set_quantizer(_A , _A , '_input_quantizer' , _A , _A ) if which in ["weight", "both"]: set_quantizer(_A , _A , '_weight_quantizer' , _A , _A ) logger.info(_A ) def lowerCAmelCase__ ( a__: List[str] , a__: List[str] , **a__: Tuple ) -> Optional[Any]: '''simple docstring''' for name, mod in model.named_modules(): if hasattr(_A , '_input_quantizer' ) or hasattr(_A , '_weight_quantizer' ): for n in names: if re.search(_A , _A ): set_quantizers(_A , _A , **_A ) elif name.endswith('_quantizer' ): for n in names: if re.search(_A , _A ): _UpperCAmelCase = F'''Warning: changing {name:{name_width}}''' for k, v in kwargs.items(): s += F''' {k}={v}''' setattr(_A , _A , _A ) logger.info(_A )
358
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available, is_vision_available, ) lowerCAmelCase__ :Tuple = {'''configuration_beit''': ['''BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BeitConfig''', '''BeitOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :List[Any] = ['''BeitFeatureExtractor'''] lowerCAmelCase__ :Optional[Any] = ['''BeitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :int = [ '''BEIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BeitForImageClassification''', '''BeitForMaskedImageModeling''', '''BeitForSemanticSegmentation''', '''BeitModel''', '''BeitPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ :Tuple = [ '''FlaxBeitForImageClassification''', '''FlaxBeitForMaskedImageModeling''', '''FlaxBeitModel''', '''FlaxBeitPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_beit import BEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, BeitConfig, BeitOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_beit import BeitFeatureExtractor from .image_processing_beit import BeitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_beit import ( BEIT_PRETRAINED_MODEL_ARCHIVE_LIST, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, BeitPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_beit import ( FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel, FlaxBeitPreTrainedModel, ) else: import sys lowerCAmelCase__ :Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
185
0
"""simple docstring""" import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): SCREAMING_SNAKE_CASE__ = "pt" elif is_tf_available(): SCREAMING_SNAKE_CASE__ = "tf" else: SCREAMING_SNAKE_CASE__ = "jax" class lowerCAmelCase_ ( lowerCAmelCase , unittest.TestCase ): """simple docstring""" _lowerCAmelCase : Any = PerceiverTokenizer _lowerCAmelCase : List[Any] = False def snake_case ( self ): """simple docstring""" super().setUp() snake_case = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def snake_case ( self ): """simple docstring""" return PerceiverTokenizer.from_pretrained('deepmind/language-perceiver' ) def snake_case ( self , **lowerCAmelCase ): """simple docstring""" return self.tokenizer_class.from_pretrained(self.tmpdirname , **lowerCAmelCase ) def snake_case ( self , lowerCAmelCase , lowerCAmelCase=False , lowerCAmelCase=20 , lowerCAmelCase=5 ): """simple docstring""" snake_case = [] for i in range(len(lowerCAmelCase ) ): try: snake_case = tokenizer.decode([i] , clean_up_tokenization_spaces=lowerCAmelCase ) except UnicodeDecodeError: pass toks.append((i, tok) ) snake_case = list(filter(lambda lowerCAmelCase : re.match(R'^[ a-zA-Z]+$' , t[1] ) , lowerCAmelCase ) ) snake_case = list(filter(lambda lowerCAmelCase : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=lowerCAmelCase ) , lowerCAmelCase ) ) if max_length is not None and len(lowerCAmelCase ) > max_length: snake_case = toks[:max_length] if min_length is not None and len(lowerCAmelCase ) < min_length and len(lowerCAmelCase ) > 0: while len(lowerCAmelCase ) < min_length: snake_case = toks + toks # toks_str = [t[1] for t in toks] snake_case = [t[0] for t in toks] # Ensure consistency snake_case = tokenizer.decode(lowerCAmelCase , clean_up_tokenization_spaces=lowerCAmelCase ) if " " not in output_txt and len(lowerCAmelCase ) > 1: snake_case = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=lowerCAmelCase ) + ' ' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=lowerCAmelCase ) ) if with_prefix_space: snake_case = ' ' + output_txt snake_case = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) return output_txt, output_ids def snake_case ( self ): """simple docstring""" snake_case = self.perceiver_tokenizer snake_case = 'Unicode €.' snake_case = tokenizer(lowerCAmelCase ) snake_case = [4, 91, 1_16, 1_11, 1_05, 1_17, 1_06, 1_07, 38, 2_32, 1_36, 1_78, 52, 5] self.assertEqual(encoded['input_ids'] , lowerCAmelCase ) # decoding snake_case = tokenizer.decode(lowerCAmelCase ) self.assertEqual(lowerCAmelCase , '[CLS]Unicode €.[SEP]' ) snake_case = tokenizer('e è é ê ë' ) snake_case = [4, 1_07, 38, 2_01, 1_74, 38, 2_01, 1_75, 38, 2_01, 1_76, 38, 2_01, 1_77, 5] self.assertEqual(encoded['input_ids'] , lowerCAmelCase ) # decoding snake_case = tokenizer.decode(lowerCAmelCase ) self.assertEqual(lowerCAmelCase , '[CLS]e è é ê ë[SEP]' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('e è é ê ë' ) ) , '[CLS]e è é ê ë[SEP]' ) def snake_case ( self ): """simple docstring""" snake_case = self.perceiver_tokenizer snake_case = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off snake_case = [4, 71, 38, 1_14, 1_17, 1_16, 1_09, 38, 1_18, 1_03, 1_20, 1_03, 1_09, 1_20, 1_03, 1_18, 1_10, 38, 1_08, 1_17, 1_20, 38, 1_21, 1_23, 1_15, 1_15, 1_03, 1_20, 1_11, 1_28, 1_03, 1_22, 1_11, 1_17, 1_16, 52, 5, 0] # fmt: on snake_case = tokenizer(lowerCAmelCase , padding=lowerCAmelCase , return_tensors=lowerCAmelCase ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase ) if FRAMEWORK != "jax": snake_case = list(batch.input_ids.numpy()[0] ) else: snake_case = list(batch.input_ids.tolist()[0] ) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) self.assertEqual((2, 38) , batch.input_ids.shape ) self.assertEqual((2, 38) , batch.attention_mask.shape ) def snake_case ( self ): """simple docstring""" snake_case = self.perceiver_tokenizer snake_case = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] snake_case = tokenizer(lowerCAmelCase , padding=lowerCAmelCase , return_tensors=lowerCAmelCase ) # check if input_ids are returned and no decoder_input_ids self.assertIn('input_ids' , lowerCAmelCase ) self.assertIn('attention_mask' , lowerCAmelCase ) self.assertNotIn('decoder_input_ids' , lowerCAmelCase ) self.assertNotIn('decoder_attention_mask' , lowerCAmelCase ) def snake_case ( self ): """simple docstring""" snake_case = self.perceiver_tokenizer snake_case = [ 'Summary of the text.', 'Another summary.', ] snake_case = tokenizer( text_target=lowerCAmelCase , max_length=32 , padding='max_length' , truncation=lowerCAmelCase , return_tensors=lowerCAmelCase ) self.assertEqual(32 , targets['input_ids'].shape[1] ) def snake_case ( self ): """simple docstring""" snake_case = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test snake_case = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc snake_case = tempfile.mkdtemp() snake_case = ' He is very happy, UNwant\u00E9d,running' snake_case = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) tokenizer.save_pretrained(lowerCAmelCase ) snake_case = tokenizer.__class__.from_pretrained(lowerCAmelCase ) snake_case = after_tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) shutil.rmtree(lowerCAmelCase ) snake_case = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): # Isolate this from the other tests because we save additional tokens/etc snake_case = tempfile.mkdtemp() snake_case = ' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['bim', 'bambam'] ) snake_case = tokenizer.additional_special_tokens additional_special_tokens.append('new_additional_special_token' ) tokenizer.add_special_tokens({'additional_special_tokens': additional_special_tokens} ) snake_case = tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) tokenizer.save_pretrained(lowerCAmelCase ) snake_case = tokenizer.__class__.from_pretrained(lowerCAmelCase ) snake_case = after_tokenizer.encode(lowerCAmelCase , add_special_tokens=lowerCAmelCase ) self.assertListEqual(lowerCAmelCase , lowerCAmelCase ) self.assertIn('new_additional_special_token' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) snake_case = tokenizer.__class__.from_pretrained(lowerCAmelCase , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(lowerCAmelCase ) def snake_case ( self ): """simple docstring""" snake_case = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(lowerCAmelCase ) with open(os.path.join(lowerCAmelCase , 'special_tokens_map.json' ) , encoding='utf-8' ) as json_file: snake_case = json.load(lowerCAmelCase ) with open(os.path.join(lowerCAmelCase , 'tokenizer_config.json' ) , encoding='utf-8' ) as json_file: snake_case = json.load(lowerCAmelCase ) snake_case = [F"""<extra_id_{i}>""" for i in range(1_25 )] snake_case = added_tokens_extra_ids + [ 'an_additional_special_token' ] snake_case = added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(lowerCAmelCase , 'special_tokens_map.json' ) , 'w' , encoding='utf-8' ) as outfile: json.dump(lowerCAmelCase , lowerCAmelCase ) with open(os.path.join(lowerCAmelCase , 'tokenizer_config.json' ) , 'w' , encoding='utf-8' ) as outfile: json.dump(lowerCAmelCase , lowerCAmelCase ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files snake_case = tokenizer_class.from_pretrained( lowerCAmelCase , ) self.assertIn( 'an_additional_special_token' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['an_additional_special_token'] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['an_additional_special_token'] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained snake_case = added_tokens_extra_ids + [AddedToken('a_new_additional_special_token' , lstrip=lowerCAmelCase )] snake_case = tokenizer_class.from_pretrained( lowerCAmelCase , additional_special_tokens=lowerCAmelCase , ) self.assertIn('a_new_additional_special_token' , tokenizer.additional_special_tokens ) self.assertEqual( ['a_new_additional_special_token'] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['a_new_additional_special_token'] ) ) , ) def snake_case ( self ): """simple docstring""" snake_case = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([1_78] ) , '�' ) def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" pass def snake_case ( self ): """simple docstring""" snake_case = self.get_tokenizers(fast=lowerCAmelCase , do_lower_case=lowerCAmelCase ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): snake_case = ['[CLS]', 't', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 's', 't', '[SEP]'] snake_case = tokenizer.convert_tokens_to_string(lowerCAmelCase ) self.assertIsInstance(lowerCAmelCase , lowerCAmelCase )
150
"""simple docstring""" import argparse import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( CLIPTokenizer, CLIPTokenizerFast, VideoMAEImageProcessor, XCLIPConfig, XCLIPModel, XCLIPProcessor, XCLIPTextConfig, XCLIPVisionConfig, ) def lowerCAmelCase__ ( _UpperCamelCase : List[Any] , _UpperCamelCase : Any ) -> Dict: """simple docstring""" snake_case = XCLIPTextConfig() # derive patch size from model name snake_case = model_name.find('patch' ) snake_case = int(model_name[start_idx + len('patch' ) : start_idx + len('patch' ) + 2] ) snake_case = XCLIPVisionConfig(patch_size=_UpperCamelCase , num_frames=_UpperCamelCase ) if "large" in model_name: snake_case = 7_6_8 snake_case = 3_0_7_2 snake_case = 1_2 snake_case = 1_0_2_4 snake_case = 4_0_9_6 snake_case = 1_6 snake_case = 2_4 snake_case = 7_6_8 snake_case = 3_0_7_2 if model_name == "xclip-large-patch14-16-frames": snake_case = 3_3_6 snake_case = XCLIPConfig.from_text_vision_configs(_UpperCamelCase , _UpperCamelCase ) if "large" in model_name: snake_case = 7_6_8 return config def lowerCAmelCase__ ( _UpperCamelCase : Dict ) -> List[Any]: """simple docstring""" if name == "token_embedding.weight": snake_case = name.replace('token_embedding.weight' , 'text_model.embeddings.token_embedding.weight' ) if name == "positional_embedding": snake_case = name.replace('positional_embedding' , 'text_model.embeddings.position_embedding.weight' ) if "ln_1" in name: snake_case = name.replace('ln_1' , 'layer_norm1' ) if "ln_2" in name: snake_case = name.replace('ln_2' , 'layer_norm2' ) if "c_fc" in name: snake_case = name.replace('c_fc' , 'fc1' ) if "c_proj" in name: snake_case = name.replace('c_proj' , 'fc2' ) if name.startswith('transformer.resblocks' ): snake_case = name.replace('transformer.resblocks' , 'text_model.encoder.layers' ) if "attn.out_proj" in name and "message" not in name: snake_case = name.replace('attn.out_proj' , 'self_attn.out_proj' ) if "ln_final" in name: snake_case = name.replace('ln_final' , 'text_model.final_layer_norm' ) # visual encoder if name == "visual.class_embedding": snake_case = name.replace('visual.class_embedding' , 'vision_model.embeddings.class_embedding' ) if name == "visual.positional_embedding": snake_case = name.replace('visual.positional_embedding' , 'vision_model.embeddings.position_embedding.weight' ) if name.startswith('visual.transformer.resblocks' ): snake_case = name.replace('visual.transformer.resblocks' , 'vision_model.encoder.layers' ) if "visual.conv1" in name: snake_case = name.replace('visual.conv1' , 'vision_model.embeddings.patch_embedding' ) if "visual.ln_pre" in name: snake_case = name.replace('visual.ln_pre' , 'vision_model.pre_layernorm' ) if "visual.ln_post" in name: snake_case = name.replace('visual.ln_post' , 'vision_model.post_layernorm' ) if "visual.proj" in name: snake_case = name.replace('visual.proj' , 'visual_projection.weight' ) if "text_projection" in name: snake_case = name.replace('text_projection' , 'text_projection.weight' ) # things on top if "prompts_visual_proj" in name: snake_case = name.replace('prompts_visual_proj' , 'prompts_visual_projection' ) if "prompts_visual_ln" in name: snake_case = name.replace('prompts_visual_ln' , 'prompts_visual_layernorm' ) # mit if name == "mit.positional_embedding": snake_case = name.replace('positional' , 'position' ) if name.startswith('mit.resblocks' ): snake_case = name.replace('mit.resblocks' , 'mit.encoder.layers' ) # prompts generator if name.startswith('prompts_generator.norm' ): snake_case = name.replace('prompts_generator.norm' , 'prompts_generator.layernorm' ) return name def lowerCAmelCase__ ( _UpperCamelCase : List[Any] , _UpperCamelCase : List[str] ) -> List[str]: """simple docstring""" for key in orig_state_dict.copy().keys(): snake_case = orig_state_dict.pop(_UpperCamelCase ) if "attn.in_proj" in key: snake_case = key.split('.' ) if key.startswith('visual' ): snake_case = key_split[3] snake_case = config.vision_config.hidden_size if "message_attn" in key: if "weight" in key: snake_case = val[ :dim, : ] snake_case = val[ dim : dim * 2, : ] snake_case = val[ -dim:, : ] else: snake_case = val[ :dim ] snake_case = val[ dim : dim * 2 ] snake_case = val[ -dim: ] else: if "weight" in key: snake_case = val[ :dim, : ] snake_case = val[ dim : dim * 2, : ] snake_case = val[ -dim:, : ] else: snake_case = val[:dim] snake_case = val[ dim : dim * 2 ] snake_case = val[-dim:] elif key.startswith('mit' ): snake_case = key_split[2] snake_case = config.vision_config.mit_hidden_size if "weight" in key: snake_case = val[:dim, :] snake_case = val[dim : dim * 2, :] snake_case = val[-dim:, :] else: snake_case = val[:dim] snake_case = val[dim : dim * 2] snake_case = val[-dim:] else: snake_case = key_split[2] snake_case = config.text_config.hidden_size if "weight" in key: snake_case = val[:dim, :] snake_case = val[ dim : dim * 2, : ] snake_case = val[-dim:, :] else: snake_case = val[:dim] snake_case = val[ dim : dim * 2 ] snake_case = val[-dim:] else: snake_case = rename_key(_UpperCamelCase ) if new_key_name in ["visual_projection.weight", "text_projection.weight"]: snake_case = val.T snake_case = val return orig_state_dict def lowerCAmelCase__ ( _UpperCamelCase : List[str] ) -> Optional[Any]: """simple docstring""" if num_frames == 8: snake_case = 'eating_spaghetti_8_frames.npy' elif num_frames == 1_6: snake_case = 'eating_spaghetti.npy' elif num_frames == 3_2: snake_case = 'eating_spaghetti_32_frames.npy' snake_case = hf_hub_download( repo_id='hf-internal-testing/spaghetti-video' , filename=_UpperCamelCase , repo_type='dataset' , ) snake_case = np.load(_UpperCamelCase ) return list(_UpperCamelCase ) def lowerCAmelCase__ ( _UpperCamelCase : str , _UpperCamelCase : Tuple=None , _UpperCamelCase : Union[str, Any]=False ) -> Optional[int]: """simple docstring""" snake_case = { # fully supervised kinetics-400 checkpoints 'xclip-base-patch32': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_8.pth', 'xclip-base-patch32-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_32_16.pth' ), 'xclip-base-patch16': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_8.pth', 'xclip-base-patch16-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k400_16_16.pth' ), 'xclip-large-patch14': 'https://drive.google.com/u/0/uc?id=1NUOImq0o5DlQTST17iIP3vG7DgmHQuCx&amp;export=download&amp;confirm=t&amp;uuid=b26caedc-88e2-473e-830a-9d158b653cdb', 'xclip-large-patch14-16-frames': 'https://drive.google.com/u/0/uc?id=1FOYgnJc097OJ4lGwtRCCydQyVPJEOH7d&amp;export=download&amp;confirm=t&amp;uuid=538fa810-e671-4050-b385-9a623f89804f', # fully supervised kinetics-600 checkpoints 'xclip-base-patch16-kinetics-600': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_8.pth' ), 'xclip-base-patch16-kinetics-600-16-frames': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/k600_16_16.pth' ), 'xclip-large-patch14-kinetics-600': 'https://drive.google.com/u/0/uc?id=1FV8C1INuM91sLAN4ImjzePLIlpMSihwV&amp;export=download&amp;confirm=t&amp;uuid=141d4977-4a65-44ae-864f-4b0c19f838be', # few shot 'xclip-base-patch16-hmdb-2-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_2.pth' ), 'xclip-base-patch16-hmdb-4-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_4.pth' ), 'xclip-base-patch16-hmdb-8-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_8.pth' ), 'xclip-base-patch16-hmdb-16-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_hmdb_16.pth' ), 'xclip-base-patch16-ucf-2-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_2.pth' ), 'xclip-base-patch16-ucf-4-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_4.pth' ), 'xclip-base-patch16-ucf-8-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_8.pth' ), 'xclip-base-patch16-ucf-16-shot': ( 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/few_ucf_16.pth' ), # zero shot 'xclip-base-patch16-zero-shot': 'https://github.com/nbl97/X-CLIP_Model_Zoo/releases/download/v1.0/zero.pth', } snake_case = model_to_url[model_name] snake_case = 8 if "16-frames" in model_name: snake_case = 1_6 elif "shot" in model_name: snake_case = 3_2 snake_case = get_xclip_config(_UpperCamelCase , _UpperCamelCase ) snake_case = XCLIPModel(_UpperCamelCase ) model.eval() if "drive" in checkpoint_url: snake_case = 'pytorch_model.bin' gdown.cached_download(_UpperCamelCase , _UpperCamelCase , quiet=_UpperCamelCase ) snake_case = torch.load(_UpperCamelCase , map_location='cpu' )['model'] else: snake_case = torch.hub.load_state_dict_from_url(_UpperCamelCase )['model'] snake_case = convert_state_dict(_UpperCamelCase , _UpperCamelCase ) snake_case = XCLIPModel(_UpperCamelCase ) snake_case ,snake_case = model.load_state_dict(_UpperCamelCase , strict=_UpperCamelCase ) assert missing_keys == ["text_model.embeddings.position_ids", "vision_model.embeddings.position_ids"] model.eval() snake_case = 3_3_6 if model_name == 'xclip-large-patch14-16-frames' else 2_2_4 snake_case = VideoMAEImageProcessor(size=_UpperCamelCase ) snake_case = CLIPTokenizer.from_pretrained('openai/clip-vit-base-patch32' ) snake_case = CLIPTokenizerFast.from_pretrained('openai/clip-vit-base-patch32' ) snake_case = XCLIPProcessor(image_processor=_UpperCamelCase , tokenizer=_UpperCamelCase ) snake_case = prepare_video(_UpperCamelCase ) snake_case = processor( text=['playing sports', 'eating spaghetti', 'go shopping'] , videos=_UpperCamelCase , return_tensors='pt' , padding=_UpperCamelCase ) print('Shape of pixel values:' , inputs.pixel_values.shape ) with torch.no_grad(): snake_case = model(**_UpperCamelCase ) # Verify outputs snake_case = outputs.logits_per_video snake_case = logits_per_video.softmax(dim=1 ) print('Probs:' , _UpperCamelCase ) # kinetics-400 if model_name == "xclip-base-patch32": snake_case = torch.tensor([[0.00_19, 0.99_51, 0.00_30]] ) elif model_name == "xclip-base-patch32-16-frames": snake_case = torch.tensor([[7.0_9_9_9e-0_4, 9.9_8_8_3e-0_1, 4.5_5_8_0e-0_4]] ) elif model_name == "xclip-base-patch16": snake_case = torch.tensor([[0.00_83, 0.96_81, 0.02_36]] ) elif model_name == "xclip-base-patch16-16-frames": snake_case = torch.tensor([[7.6_9_3_7e-0_4, 9.9_7_2_8e-0_1, 1.9_4_7_3e-0_3]] ) elif model_name == "xclip-large-patch14": snake_case = torch.tensor([[0.00_62, 0.98_64, 0.00_75]] ) elif model_name == "xclip-large-patch14-16-frames": snake_case = torch.tensor([[3.3_8_7_7e-0_4, 9.9_9_3_7e-0_1, 2.8_8_8_8e-0_4]] ) # kinetics-600 elif model_name == "xclip-base-patch16-kinetics-600": snake_case = torch.tensor([[0.05_55, 0.89_14, 0.05_31]] ) elif model_name == "xclip-base-patch16-kinetics-600-16-frames": snake_case = torch.tensor([[3.8_5_5_4e-0_4, 9.9_9_2_9e-0_1, 3.2_7_5_4e-0_4]] ) elif model_name == "xclip-large-patch14-kinetics-600": snake_case = torch.tensor([[0.00_36, 0.99_20, 0.00_45]] ) # few shot elif model_name == "xclip-base-patch16-hmdb-2-shot": snake_case = torch.tensor([[7.1_8_9_0e-0_6, 9.9_9_9_4e-0_1, 5.6_5_5_9e-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-4-shot": snake_case = torch.tensor([[1.0_3_2_0e-0_5, 9.9_9_9_3e-0_1, 6.2_4_3_5e-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-8-shot": snake_case = torch.tensor([[4.1_3_7_7e-0_6, 9.9_9_9_0e-0_1, 9.8_3_8_6e-0_5]] ) elif model_name == "xclip-base-patch16-hmdb-16-shot": snake_case = torch.tensor([[4.1_3_4_7e-0_5, 9.9_9_6_2e-0_1, 3.3_4_1_1e-0_4]] ) elif model_name == "xclip-base-patch16-ucf-2-shot": snake_case = torch.tensor([[8.5_8_5_7e-0_5, 9.9_9_2_8e-0_1, 6.3_2_9_1e-0_4]] ) elif model_name == "xclip-base-patch16-ucf-4-shot": snake_case = torch.tensor([[8.5_8_5_7e-0_5, 9.9_9_2_8e-0_1, 6.3_2_9_1e-0_4]] ) elif model_name == "xclip-base-patch16-ucf-8-shot": snake_case = torch.tensor([[0.00_27, 0.99_04, 0.00_70]] ) elif model_name == "xclip-base-patch16-ucf-16-shot": snake_case = torch.tensor([[9.8_2_1_9e-0_4, 9.9_5_9_3e-0_1, 3.0_8_6_3e-0_3]] ) # zero shot elif model_name == "xclip-base-patch16-zero-shot": snake_case = torch.tensor([[3.5_0_8_2e-0_4, 9.9_7_8_5e-0_1, 1.7_9_6_6e-0_3]] ) else: raise ValueError(f"""Model name {model_name} not supported""" ) assert torch.allclose(_UpperCamelCase , _UpperCamelCase , atol=1e-3 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_UpperCamelCase ) if push_to_hub: print('Pushing model, processor and slow tokenizer files to the hub...' ) model.push_to_hub(_UpperCamelCase , organization='nielsr' ) processor.push_to_hub(_UpperCamelCase , organization='nielsr' ) slow_tokenizer.push_to_hub(_UpperCamelCase , organization='nielsr' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="xclip-base-patch32", type=str, help="Name of the model.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) SCREAMING_SNAKE_CASE__ = parser.parse_args() convert_xclip_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
150
1
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" super().tearDown() gc.collect() def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : Any = FlaxStableDiffusionPipeline.from_pretrained( '''stabilityai/stable-diffusion-2''' , revision='''bf16''' , dtype=jnp.bfloataa , ) _lowerCamelCase : Tuple = '''A painting of a squirrel eating a burger''' _lowerCamelCase : Dict = jax.device_count() _lowerCamelCase : Optional[Any] = num_samples * [prompt] _lowerCamelCase : Any = sd_pipe.prepare_inputs(__lowerCAmelCase ) _lowerCamelCase : Any = replicate(__lowerCAmelCase ) _lowerCamelCase : Tuple = shard(__lowerCAmelCase ) _lowerCamelCase : List[str] = jax.random.PRNGKey(0 ) _lowerCamelCase : Tuple = jax.random.split(__lowerCAmelCase , jax.device_count() ) _lowerCamelCase : str = sd_pipe(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , num_inference_steps=2_5 , jit=__lowerCAmelCase )[0] assert images.shape == (jax.device_count(), 1, 7_6_8, 7_6_8, 3) _lowerCamelCase : Tuple = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _lowerCamelCase : Dict = images[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] _lowerCamelCase : str = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _lowerCamelCase : Optional[int] = jnp.array([0.42_38, 0.44_14, 0.43_95, 0.44_53, 0.46_29, 0.45_90, 0.45_31, 0.4_55_08, 0.45_12] ) print(f'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2 def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : List[Any] = '''stabilityai/stable-diffusion-2''' _lowerCamelCase : Optional[int] = FlaxDPMSolverMultistepScheduler.from_pretrained(__lowerCAmelCase , subfolder='''scheduler''' ) _lowerCamelCase : Dict = FlaxStableDiffusionPipeline.from_pretrained( __lowerCAmelCase , scheduler=__lowerCAmelCase , revision='''bf16''' , dtype=jnp.bfloataa , ) _lowerCamelCase : Optional[int] = scheduler_params _lowerCamelCase : Union[str, Any] = '''A painting of a squirrel eating a burger''' _lowerCamelCase : Any = jax.device_count() _lowerCamelCase : Tuple = num_samples * [prompt] _lowerCamelCase : str = sd_pipe.prepare_inputs(__lowerCAmelCase ) _lowerCamelCase : str = replicate(__lowerCAmelCase ) _lowerCamelCase : Any = shard(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = jax.random.PRNGKey(0 ) _lowerCamelCase : int = jax.random.split(__lowerCAmelCase , jax.device_count() ) _lowerCamelCase : Tuple = sd_pipe(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , num_inference_steps=2_5 , jit=__lowerCAmelCase )[0] assert images.shape == (jax.device_count(), 1, 7_6_8, 7_6_8, 3) _lowerCamelCase : Any = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) _lowerCamelCase : int = images[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] _lowerCamelCase : Dict = jnp.asarray(jax.device_get(image_slice.flatten() ) ) _lowerCamelCase : List[str] = jnp.array([0.43_36, 0.4_29_69, 0.44_53, 0.41_99, 0.42_97, 0.45_31, 0.44_34, 0.44_34, 0.42_97] ) print(f'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1E-2
368
"""simple docstring""" import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = [ '''word_embeddings_layernorm.weight''', '''word_embeddings_layernorm.bias''', '''input_layernorm.weight''', '''input_layernorm.bias''', '''post_attention_layernorm.weight''', '''post_attention_layernorm.bias''', '''self_attention.dense.bias''', '''mlp.dense_4h_to_h.bias''', '''ln_f.weight''', '''ln_f.bias''', ] lowerCAmelCase__ = [ '''mlp.dense_4h_to_h.weight''', '''self_attention.dense.weight''', ] def snake_case_ ( A_ : int, A_ : Dict ): '''simple docstring''' _lowerCamelCase : List[Any] = { '''word_embeddings.weight''': '''word_embeddings.weight''', '''word_embeddings.norm.weight''': '''word_embeddings_layernorm.weight''', '''word_embeddings.norm.bias''': '''word_embeddings_layernorm.bias''', '''weight''': '''ln_f.weight''', '''bias''': '''ln_f.bias''', } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks _lowerCamelCase : Union[str, Any] = int(re.match(R'''.*layer_(\d*).*''', A_ )[1] ) layer_number -= 3 return F'''h.{layer_number}.''' + key def snake_case_ ( A_ : List[Any] ): '''simple docstring''' if dtype == torch.bool: return 1 / 8 _lowerCamelCase : List[str] = re.search(R'''[^\d](\d+)$''', str(A_ ) ) if bit_search is None: raise ValueError(F'''`dtype` is not a valid dtype: {dtype}.''' ) _lowerCamelCase : Optional[Any] = int(bit_search.groups()[0] ) return bit_size // 8 def snake_case_ ( A_ : str, A_ : Any, A_ : int, A_ : List[str], A_ : Any ): '''simple docstring''' if bloom_config_file == "": _lowerCamelCase : Dict = BloomConfig() else: _lowerCamelCase : Any = BloomConfig.from_json_file(A_ ) if shard_model: _lowerCamelCase : Optional[int] = os.listdir(A_ ) _lowerCamelCase : List[str] = sorted(filter(lambda A_ : s.startswith('''layer''' ) and "model_00" in s, A_ ) ) _lowerCamelCase : str = {'''weight_map''': {}, '''metadata''': {}} _lowerCamelCase : List[str] = 0 _lowerCamelCase : str = None _lowerCamelCase : str = BloomConfig() for j, file in enumerate(A_ ): print('''Processing file: {}'''.format(A_ ) ) _lowerCamelCase : List[Any] = None for i in range(A_ ): # load all TP files _lowerCamelCase : Any = file.replace('''model_00''', F'''model_0{i}''' ) _lowerCamelCase : Any = torch.load(os.path.join(A_, A_ ), map_location='''cpu''' ) # Rename keys in the transformers names _lowerCamelCase : Optional[Any] = list(temp.keys() ) for key in keys: _lowerCamelCase : List[Any] = temp.pop(A_ ) if tensors is None: _lowerCamelCase : Any = temp else: for key in tensors.keys(): if any(key.endswith(A_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel _lowerCamelCase : List[Any] = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks _lowerCamelCase : Optional[Any] = torch.cat([tensors[key], temp[key]], dim=A_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(A_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): _lowerCamelCase : Optional[Any] = tensors[key] / pretraining_tp torch.save( A_, os.path.join( A_, '''pytorch_model_{}-of-{}.bin'''.format(str(j + 1 ).zfill(5 ), str(len(A_ ) ).zfill(5 ) ), ), ) for key in tensors.keys(): _lowerCamelCase : str = tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: _lowerCamelCase : Tuple = '''pytorch_model_{}-of-{}.bin'''.format( str(j + 1 ).zfill(5 ), str(len(A_ ) ).zfill(5 ) ) _lowerCamelCase : List[Any] = BloomConfig() _lowerCamelCase : Tuple = pytorch_dump_folder_path + '''/''' + CONFIG_NAME _lowerCamelCase : Union[str, Any] = total_size with open(A_, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) with open(os.path.join(A_, WEIGHTS_NAME + '''.index.json''' ), '''w''', encoding='''utf-8''' ) as f: _lowerCamelCase : Any = json.dumps(A_, indent=2, sort_keys=A_ ) + '''\n''' f.write(A_ ) else: _lowerCamelCase : Tuple = BloomModel(A_ ) _lowerCamelCase : Optional[int] = os.listdir(A_ ) _lowerCamelCase : Union[str, Any] = sorted(filter(lambda A_ : s.startswith('''layer''' ) and "model_00" in s, A_ ) ) _lowerCamelCase : int = None for i, file in enumerate(A_ ): _lowerCamelCase : Optional[int] = None for i in range(A_ ): # load all TP files _lowerCamelCase : str = file.replace('''model_00''', F'''model_0{i}''' ) _lowerCamelCase : List[Any] = torch.load(os.path.join(A_, A_ ), map_location='''cpu''' ) # Rename keys in the transformers names _lowerCamelCase : List[Any] = list(temp.keys() ) for key in keys: _lowerCamelCase : Dict = temp.pop(A_ ) if tensors is None: _lowerCamelCase : int = temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(A_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel _lowerCamelCase : Dict = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks _lowerCamelCase : Optional[Any] = torch.cat([tensors[key], temp[key]], dim=A_ ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(A_ ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): _lowerCamelCase : List[Any] = tensors[key] / pretraining_tp _lowerCamelCase : List[str] = model.load_state_dict(A_, strict=A_ ) assert not other_keys.unexpected_keys, F'''The keys {other_keys.unexpected_keys} are unexpected''' if missing_keys is None: _lowerCamelCase : Optional[Any] = set(other_keys.missing_keys ) else: _lowerCamelCase : Union[str, Any] = missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, F'''The keys {missing_keys} are missing''' # Save pytorch-model os.makedirs(A_, exist_ok=A_ ) _lowerCamelCase : Union[str, Any] = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME _lowerCamelCase : Tuple = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}''' ) if config.torch_dtype is not None: _lowerCamelCase : Dict = model.to(config.torch_dtype ) torch.save(model.state_dict(), A_ ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(A_, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bloom_checkpoint_path''', default=None, type=str, required=True, help='''Path to the Megatron-LM checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--bloom_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--shard_model''', action='''store_true''', help='''An optional setting to shard the output model \nThis enables sharding the converted checkpoint''', ) parser.add_argument( '''--pretraining_tp''', default=4, type=int, help='''Pretraining TP rank that has been used when training the model in Megatron-LM \n''', ) lowerCAmelCase__ = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
175
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer _lowerCamelCase : str = logging.get_logger(__name__) _lowerCamelCase : Optional[int] = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} _lowerCamelCase : Optional[int] = { """vocab_file""": { """junnyu/roformer_chinese_small""": """https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt""", """junnyu/roformer_chinese_base""": """https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt""", """junnyu/roformer_chinese_char_small""": ( """https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt""" ), """junnyu/roformer_chinese_char_base""": ( """https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt""" ), """junnyu/roformer_small_discriminator""": ( """https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt""" ), """junnyu/roformer_small_generator""": ( """https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt""" ), } } _lowerCamelCase : Optional[Any] = { """junnyu/roformer_chinese_small""": 1536, """junnyu/roformer_chinese_base""": 1536, """junnyu/roformer_chinese_char_small""": 512, """junnyu/roformer_chinese_char_base""": 512, """junnyu/roformer_small_discriminator""": 128, """junnyu/roformer_small_generator""": 128, } _lowerCamelCase : List[Any] = { """junnyu/roformer_chinese_small""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_base""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_char_small""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_char_base""": {"""do_lower_case""": True}, """junnyu/roformer_small_discriminator""": {"""do_lower_case""": True}, """junnyu/roformer_small_generator""": {"""do_lower_case""": True}, } class UpperCamelCase_ ( UpperCAmelCase__ ): '''simple docstring''' UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = RoFormerTokenizer def __init__( self : Any , UpperCAmelCase__ : int=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : str=True , UpperCAmelCase__ : Optional[int]="[UNK]" , UpperCAmelCase__ : Optional[int]="[SEP]" , UpperCAmelCase__ : Union[str, Any]="[PAD]" , UpperCAmelCase__ : str="[CLS]" , UpperCAmelCase__ : Tuple="[MASK]" , UpperCAmelCase__ : int=True , UpperCAmelCase__ : Tuple=None , **UpperCAmelCase__ : Optional[Any] , ) ->List[Any]: '''simple docstring''' super().__init__( UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , do_lower_case=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , tokenize_chinese_chars=UpperCAmelCase__ , strip_accents=UpperCAmelCase__ , **UpperCAmelCase__ , ) A__ = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( pre_tok_state.get('''lowercase''' , UpperCAmelCase__) != do_lower_case or pre_tok_state.get('''strip_accents''' , UpperCAmelCase__) != strip_accents ): A__ = getattr(UpperCAmelCase__ , pre_tok_state.pop('''type''')) A__ = do_lower_case A__ = strip_accents A__ = pre_tok_class(**UpperCAmelCase__) A__ = do_lower_case def __getstate__( self : Tuple) ->List[str]: '''simple docstring''' A__ = self.__dict__.copy() A__ = BertPreTokenizer() return state def __setstate__( self : Any , UpperCAmelCase__ : Tuple) ->List[Any]: '''simple docstring''' A__ = d A__ = self.__dict__['''_tokenizer'''].get_vocab() A__ = PreTokenizer.custom(JiebaPreTokenizer(UpperCAmelCase__)) def SCREAMING_SNAKE_CASE ( self : Any , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Union[str, Any]=None) ->Dict: '''simple docstring''' A__ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def SCREAMING_SNAKE_CASE ( self : Dict , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None) ->List[int]: '''simple docstring''' A__ = [self.sep_token_id] A__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def SCREAMING_SNAKE_CASE ( self : Tuple , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None) ->Tuple[str]: '''simple docstring''' A__ = self._tokenizer.model.save(UpperCAmelCase__ , name=UpperCAmelCase__) return tuple(UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Any , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : List[Any]=False , **UpperCAmelCase__ : str , ) ->List[str]: '''simple docstring''' A__ = BertPreTokenizer() return super().save_pretrained(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , **UpperCAmelCase__)
14
import os import sys import unittest _lowerCamelCase : Optional[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import get_test_info # noqa: E402 from get_test_info import ( # noqa: E402 get_model_to_test_mapping, get_model_to_tester_mapping, get_test_to_tester_mapping, ) _lowerCamelCase : Any = os.path.join("""tests""", """models""", """bert""", """test_modeling_bert.py""") _lowerCamelCase : str = os.path.join("""tests""", """models""", """blip""", """test_modeling_blip.py""") class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE ( self : Tuple) ->Tuple: '''simple docstring''' A__ = get_test_to_tester_mapping(UpperCAmelCase__) A__ = get_test_to_tester_mapping(UpperCAmelCase__) A__ = {'''BertModelTest''': '''BertModelTester'''} A__ = { '''BlipModelTest''': '''BlipModelTester''', '''BlipTextImageModelTest''': '''BlipTextImageModelsModelTester''', '''BlipTextModelTest''': '''BlipTextModelTester''', '''BlipTextRetrievalModelTest''': '''BlipTextRetrievalModelTester''', '''BlipVQAModelTest''': '''BlipVQAModelTester''', '''BlipVisionModelTest''': '''BlipVisionModelTester''', } self.assertEqual(get_test_info.to_json(UpperCAmelCase__) , UpperCAmelCase__) self.assertEqual(get_test_info.to_json(UpperCAmelCase__) , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Tuple) ->List[Any]: '''simple docstring''' A__ = get_model_to_test_mapping(UpperCAmelCase__) A__ = get_model_to_test_mapping(UpperCAmelCase__) A__ = { '''BertForMaskedLM''': ['''BertModelTest'''], '''BertForMultipleChoice''': ['''BertModelTest'''], '''BertForNextSentencePrediction''': ['''BertModelTest'''], '''BertForPreTraining''': ['''BertModelTest'''], '''BertForQuestionAnswering''': ['''BertModelTest'''], '''BertForSequenceClassification''': ['''BertModelTest'''], '''BertForTokenClassification''': ['''BertModelTest'''], '''BertLMHeadModel''': ['''BertModelTest'''], '''BertModel''': ['''BertModelTest'''], } A__ = { '''BlipForConditionalGeneration''': ['''BlipTextImageModelTest'''], '''BlipForImageTextRetrieval''': ['''BlipTextRetrievalModelTest'''], '''BlipForQuestionAnswering''': ['''BlipVQAModelTest'''], '''BlipModel''': ['''BlipModelTest'''], '''BlipTextModel''': ['''BlipTextModelTest'''], '''BlipVisionModel''': ['''BlipVisionModelTest'''], } self.assertEqual(get_test_info.to_json(UpperCAmelCase__) , UpperCAmelCase__) self.assertEqual(get_test_info.to_json(UpperCAmelCase__) , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->str: '''simple docstring''' A__ = get_model_to_tester_mapping(UpperCAmelCase__) A__ = get_model_to_tester_mapping(UpperCAmelCase__) A__ = { '''BertForMaskedLM''': ['''BertModelTester'''], '''BertForMultipleChoice''': ['''BertModelTester'''], '''BertForNextSentencePrediction''': ['''BertModelTester'''], '''BertForPreTraining''': ['''BertModelTester'''], '''BertForQuestionAnswering''': ['''BertModelTester'''], '''BertForSequenceClassification''': ['''BertModelTester'''], '''BertForTokenClassification''': ['''BertModelTester'''], '''BertLMHeadModel''': ['''BertModelTester'''], '''BertModel''': ['''BertModelTester'''], } A__ = { '''BlipForConditionalGeneration''': ['''BlipTextImageModelsModelTester'''], '''BlipForImageTextRetrieval''': ['''BlipTextRetrievalModelTester'''], '''BlipForQuestionAnswering''': ['''BlipVQAModelTester'''], '''BlipModel''': ['''BlipModelTester'''], '''BlipTextModel''': ['''BlipTextModelTester'''], '''BlipVisionModel''': ['''BlipVisionModelTester'''], } self.assertEqual(get_test_info.to_json(UpperCAmelCase__) , UpperCAmelCase__) self.assertEqual(get_test_info.to_json(UpperCAmelCase__) , UpperCAmelCase__)
14
1
"""simple docstring""" import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.utils import logging __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): super().__init__() if safety_checker is None: logger.warning( F"""You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure""" ''' that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered''' ''' results in services or applications open to the public. Both the diffusers team and Hugging Face''' ''' strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling''' ''' it only for use-cases that involve analyzing network behavior or auditing its results. For more''' ''' information, please have a look at https://github.com/huggingface/diffusers/pull/254 .''' ) self.register_modules( speech_model=lowerCamelCase__ , speech_processor=lowerCamelCase__ , vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , unet=lowerCamelCase__ , scheduler=lowerCamelCase__ , feature_extractor=lowerCamelCase__ , ) def snake_case__ ( self , lowerCamelCase__ = "auto" ): if slice_size == "auto": _lowerCamelCase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase__ ) def snake_case__ ( self ): self.enable_attention_slicing(lowerCamelCase__ ) @torch.no_grad() def __call__( self , lowerCamelCase__ , lowerCamelCase__=1_6_0_0_0 , lowerCamelCase__ = 5_1_2 , lowerCamelCase__ = 5_1_2 , lowerCamelCase__ = 5_0 , lowerCamelCase__ = 7.5 , lowerCamelCase__ = None , lowerCamelCase__ = 1 , lowerCamelCase__ = 0.0 , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = "pil" , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = 1 , **lowerCamelCase__ , ): _lowerCamelCase = self.speech_processor.feature_extractor( lowerCamelCase__ , return_tensors='''pt''' , sampling_rate=lowerCamelCase__ ).input_features.to(self.device ) _lowerCamelCase = self.speech_model.generate(lowerCamelCase__ , max_length=4_8_0_0_0_0 ) _lowerCamelCase = self.speech_processor.tokenizer.batch_decode(lowerCamelCase__ , skip_special_tokens=lowerCamelCase__ , normalize=lowerCamelCase__ )[ 0 ] if isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = 1 elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = len(lowerCamelCase__ ) else: raise ValueError(F"""`prompt` has to be of type `str` or `list` but is {type(lowerCamelCase__ )}""" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F"""`height` and `width` have to be divisible by 8 but are {height} and {width}.""" ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(lowerCamelCase__ , lowerCamelCase__ ) or callback_steps <= 0) ): raise ValueError( F"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" F""" {type(lowerCamelCase__ )}.""" ) # get prompt text embeddings _lowerCamelCase = self.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=self.tokenizer.model_max_length , return_tensors='''pt''' , ) _lowerCamelCase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: _lowerCamelCase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( '''The following part of your input was truncated because CLIP can only handle sequences up to''' F""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) _lowerCamelCase = text_input_ids[:, : self.tokenizer.model_max_length] _lowerCamelCase = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = text_embeddings.shape _lowerCamelCase = text_embeddings.repeat(1 , lowerCamelCase__ , 1 ) _lowerCamelCase = text_embeddings.view(bs_embed * num_images_per_prompt , lowerCamelCase__ , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. _lowerCamelCase = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: _lowerCamelCase = 42 if negative_prompt is None: _lowerCamelCase = [''''''] * batch_size elif type(lowerCamelCase__ ) is not type(lowerCamelCase__ ): raise TypeError( F"""`negative_prompt` should be the same type to `prompt`, but got {type(lowerCamelCase__ )} !=""" F""" {type(lowerCamelCase__ )}.""" ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = [negative_prompt] elif batch_size != len(lowerCamelCase__ ): raise ValueError( F"""`negative_prompt`: {negative_prompt} has batch size {len(lowerCamelCase__ )}, but `prompt`:""" F""" {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches""" ''' the batch size of `prompt`.''' ) else: _lowerCamelCase = negative_prompt _lowerCamelCase = text_input_ids.shape[-1] _lowerCamelCase = self.tokenizer( lowerCamelCase__ , padding='''max_length''' , max_length=lowerCamelCase__ , truncation=lowerCamelCase__ , return_tensors='''pt''' , ) _lowerCamelCase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method _lowerCamelCase = uncond_embeddings.shape[1] _lowerCamelCase = uncond_embeddings.repeat(1 , lowerCamelCase__ , 1 ) _lowerCamelCase = uncond_embeddings.view(batch_size * num_images_per_prompt , lowerCamelCase__ , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _lowerCamelCase = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. _lowerCamelCase = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) _lowerCamelCase = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps _lowerCamelCase = torch.randn(lowerCamelCase__ , generator=lowerCamelCase__ , device='''cpu''' , dtype=lowerCamelCase__ ).to( self.device ) else: _lowerCamelCase = torch.randn(lowerCamelCase__ , generator=lowerCamelCase__ , device=self.device , dtype=lowerCamelCase__ ) else: if latents.shape != latents_shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) _lowerCamelCase = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(lowerCamelCase__ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand _lowerCamelCase = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler _lowerCamelCase = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] _lowerCamelCase = '''eta''' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) _lowerCamelCase = {} if accepts_eta: _lowerCamelCase = eta for i, t in enumerate(self.progress_bar(lowerCamelCase__ ) ): # expand the latents if we are doing classifier free guidance _lowerCamelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents _lowerCamelCase = self.scheduler.scale_model_input(lowerCamelCase__ , lowerCamelCase__ ) # predict the noise residual _lowerCamelCase = self.unet(lowerCamelCase__ , lowerCamelCase__ , encoder_hidden_states=lowerCamelCase__ ).sample # perform guidance if do_classifier_free_guidance: _lowerCamelCase , _lowerCamelCase = noise_pred.chunk(2 ) _lowerCamelCase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 _lowerCamelCase = self.scheduler.step(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = 1 / 0.1_8_2_1_5 * latents _lowerCamelCase = self.vae.decode(lowerCamelCase__ ).sample _lowerCamelCase = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 _lowerCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": _lowerCamelCase = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return image return StableDiffusionPipelineOutput(images=lowerCamelCase__ , nsfw_content_detected=lowerCamelCase__ )
73
"""simple docstring""" import os from collections.abc import Iterator def lowerCAmelCase_( lowercase_ : str = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(lowercase_ ): _lowerCamelCase = [d for d in dir_names if d != '''scripts''' and d[0] not in '''._'''] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(lowercase_ )[1] in (".py", ".ipynb"): yield os.path.join(lowercase_ , lowercase_ ).lstrip('''./''' ) def lowerCAmelCase_( lowercase_ : Dict ) -> List[Any]: return F"""{i * " "}*""" if i else "\n##" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> str: _lowerCamelCase = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(lowercase_ ) or old_parts[i] != new_part) and new_part: print(F"""{md_prefix(lowercase_ )} {new_part.replace("_" , " " ).title()}""" ) return new_path def lowerCAmelCase_( lowercase_ : str = "." ) -> None: _lowerCamelCase = '''''' for filepath in sorted(good_file_paths(lowercase_ ) ): _lowerCamelCase , _lowerCamelCase = os.path.split(lowercase_ ) if filepath != old_path: _lowerCamelCase = print_path(lowercase_ , lowercase_ ) _lowerCamelCase = (filepath.count(os.sep ) + 1) if filepath else 0 _lowerCamelCase = F"""{filepath}/{filename}""".replace(''' ''' , '''%20''' ) _lowerCamelCase = os.path.splitext(filename.replace('''_''' , ''' ''' ).title() )[0] print(F"""{md_prefix(lowercase_ )} [{filename}]({url})""" ) if __name__ == "__main__": print_directory_md('''.''')
73
1
'''simple docstring''' from __future__ import annotations def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): lowerCamelCase__ , lowerCamelCase__ : List[Any] = array[indexa], array[indexa] def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: if length > 1: lowerCamelCase__ : Optional[Any] = int(length / 2 ) for i in range(UpperCamelCase , low + middle ): comp_and_swap(UpperCamelCase , UpperCamelCase , i + middle , UpperCamelCase ) bitonic_merge(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) bitonic_merge(UpperCamelCase , low + middle , UpperCamelCase , UpperCamelCase ) def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> None: if length > 1: lowerCamelCase__ : Tuple = int(length / 2 ) bitonic_sort(UpperCamelCase , UpperCamelCase , UpperCamelCase , 1 ) bitonic_sort(UpperCamelCase , low + middle , UpperCamelCase , 0 ) bitonic_merge(UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": _A : List[str] =input('''Enter numbers separated by a comma:\n''').strip() _A : List[str] =[int(item.strip()) for item in user_input.split(''',''')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('''\nSorted array in ascending order is: ''', end='''''') print(*unsorted, sep=''', ''') bitonic_merge(unsorted, 0, len(unsorted), 0) print('''Sorted array in descending order is: ''', end='''''') print(*unsorted, sep=''', ''')
41
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ ) -> int: if exponent == 1: return base if exponent % 2 == 0: A__ = _modexpt(lowercase_ , exponent // 2 , lowercase_ ) % modulo_value return (x * x) % modulo_value else: return (base * _modexpt(lowercase_ , exponent - 1 , lowercase_ )) % modulo_value def _SCREAMING_SNAKE_CASE ( lowercase_ = 17_77 , lowercase_ = 18_55 , lowercase_ = 8 ) -> int: A__ = base for _ in range(1 , lowercase_ ): A__ = _modexpt(lowercase_ , lowercase_ , 10**digits ) return result if __name__ == "__main__": print(f'{solution() = }')
247
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ ={ 'configuration_luke': ['LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LukeConfig'], 'tokenization_luke': ['LukeTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ =[ 'LUKE_PRETRAINED_MODEL_ARCHIVE_LIST', 'LukeForEntityClassification', 'LukeForEntityPairClassification', 'LukeForEntitySpanClassification', 'LukeForMultipleChoice', 'LukeForQuestionAnswering', 'LukeForSequenceClassification', 'LukeForTokenClassification', 'LukeForMaskedLM', 'LukeModel', 'LukePreTrainedModel', ] if TYPE_CHECKING: from .configuration_luke import LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP, LukeConfig from .tokenization_luke import LukeTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_luke import ( LUKE_PRETRAINED_MODEL_ARCHIVE_LIST, LukeForEntityClassification, LukeForEntityPairClassification, LukeForEntitySpanClassification, LukeForMaskedLM, LukeForMultipleChoice, LukeForQuestionAnswering, LukeForSequenceClassification, LukeForTokenClassification, LukeModel, LukePreTrainedModel, ) else: import sys UpperCamelCase__ =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
325
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ =np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ =[0, 25, 50] UpperCamelCase__ =[25, 50, 75] UpperCamelCase__ =fuzz.membership.trimf(X, abca) UpperCamelCase__ =fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ =np.ones(75) UpperCamelCase__ =np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ =fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ =fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ =fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ =fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ =young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ =young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ =fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ =fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
325
1
'''simple docstring''' import argparse import json from tqdm import tqdm def UpperCamelCase__ ( ): """simple docstring""" _lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--src_path""" , type=lowerCAmelCase , default="""biencoder-nq-dev.json""" , help="""Path to raw DPR training data""" , ) parser.add_argument( """--evaluation_set""" , type=lowerCAmelCase , help="""where to store parsed evaluation_set file""" , ) parser.add_argument( """--gold_data_path""" , type=lowerCAmelCase , help="""where to store parsed gold_data_path file""" , ) _lowerCAmelCase = parser.parse_args() with open(args.src_path , """r""" ) as src_file, open(args.evaluation_set , """w""" ) as eval_file, open( args.gold_data_path , """w""" ) as gold_file: _lowerCAmelCase = json.load(lowerCAmelCase ) for dpr_record in tqdm(lowerCAmelCase ): _lowerCAmelCase = dpr_record["""question"""] _lowerCAmelCase = [context["""title"""] for context in dpr_record["""positive_ctxs"""]] eval_file.write(question + """\n""" ) gold_file.write("""\t""".join(lowerCAmelCase ) + """\n""" ) if __name__ == "__main__": main()
70
'''simple docstring''' # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class UpperCAmelCase ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Optional[int] = StableDiffusionControlNetImgaImgPipeline _lowercase: Tuple = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Tuple = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'''control_image'''} ) _lowercase: Union[str, Any] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase__ ( self : List[str] ) -> List[str]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Any , __snake_case : str , __snake_case : Any=0 ) -> str: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ) _lowerCAmelCase = floats_tensor(control_image.shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : Optional[int] ) -> List[Any]: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : Tuple ) -> Optional[int]: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) class UpperCAmelCase ( snake_case_ , snake_case_ , unittest.TestCase ): _lowercase: Any = StableDiffusionControlNetImgaImgPipeline _lowercase: Dict = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} _lowercase: List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS _lowercase: Any = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def lowercase__ ( self : Optional[Any] ) -> Union[str, Any]: torch.manual_seed(0 ) _lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) torch.manual_seed(0 ) def init_weights(__snake_case : Optional[Any] ): if isinstance(__snake_case , torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = ControlNetModel( block_out_channels=(32, 64) , layers_per_block=2 , in_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , cross_attention_dim=32 , conditioning_embedding_out_channels=(16, 32) , ) controlneta.controlnet_down_blocks.apply(__snake_case ) torch.manual_seed(0 ) _lowerCAmelCase = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="""scaled_linear""" , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) _lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) _lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) _lowerCAmelCase = CLIPTextModel(__snake_case ) _lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) _lowerCAmelCase = MultiControlNetModel([controlneta, controlneta] ) _lowerCAmelCase = { """unet""": unet, """controlnet""": controlnet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowercase__ ( self : Tuple , __snake_case : int , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith("""mps""" ): _lowerCAmelCase = torch.manual_seed(__snake_case ) else: _lowerCAmelCase = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) _lowerCAmelCase = 2 _lowerCAmelCase = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) , generator=__snake_case , device=torch.device(__snake_case ) , ), ] _lowerCAmelCase = floats_tensor(control_image[0].shape , rng=random.Random(__snake_case ) ).to(__snake_case ) _lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 )[0] _lowerCAmelCase = Image.fromarray(np.uinta(__snake_case ) ).convert("""RGB""" ).resize((64, 64) ) _lowerCAmelCase = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", """image""": image, """control_image""": control_image, } return inputs def lowercase__ ( self : List[str] ) -> Dict: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) _lowerCAmelCase = 10.0 _lowerCAmelCase = 4 _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.1 , control_guidance_end=0.2 )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7] )[0] _lowerCAmelCase = self.get_dummy_inputs(__snake_case ) _lowerCAmelCase = steps _lowerCAmelCase = scale _lowerCAmelCase = pipe(**__snake_case , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 assert np.sum(np.abs(output_a - output_a ) ) > 1E-3 def lowercase__ ( self : int ) -> str: return self._test_attention_slicing_forward_pass(expected_max_diff=2E-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def lowercase__ ( self : Optional[Any] ) -> Dict: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2E-3 ) def lowercase__ ( self : int ) -> str: self._test_inference_batch_single_identical(expected_max_diff=2E-3 ) def lowercase__ ( self : Union[str, Any] ) -> Optional[Any]: _lowerCAmelCase = self.get_dummy_components() _lowerCAmelCase = self.pipeline_class(**__snake_case ) pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(__snake_case ) except NotImplementedError: pass @slow @require_torch_gpu class UpperCAmelCase ( unittest.TestCase ): def lowercase__ ( self : Union[str, Any] ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase__ ( self : List[str] ) -> Any: _lowerCAmelCase = ControlNetModel.from_pretrained("""lllyasviel/sd-controlnet-canny""" ) _lowerCAmelCase = StableDiffusionControlNetImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , safety_checker=__snake_case , controlnet=__snake_case ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=__snake_case ) _lowerCAmelCase = torch.Generator(device="""cpu""" ).manual_seed(0 ) _lowerCAmelCase = """evil space-punk bird""" _lowerCAmelCase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = load_image( """https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png""" ).resize((5_12, 5_12) ) _lowerCAmelCase = pipe( __snake_case , __snake_case , control_image=__snake_case , generator=__snake_case , output_type="""np""" , num_inference_steps=50 , strength=0.6 , ) _lowerCAmelCase = output.images[0] assert image.shape == (5_12, 5_12, 3) _lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy""" ) assert np.abs(expected_image - image ).max() < 9E-2
70
1
"""simple docstring""" import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaskFormerForInstanceSegmentation, MaskFormerModel if is_vision_available(): from transformers import MaskFormerImageProcessor if is_vision_available(): from PIL import Image class __a : def __init__( self , a__ , a__=2 , a__=True , a__=False , a__=10 , a__=3 , a__=32 * 4 , a__=32 * 6 , a__=4 , a__=32 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = is_training _lowerCamelCase = use_auxiliary_loss _lowerCamelCase = num_queries _lowerCamelCase = num_channels _lowerCamelCase = min_size _lowerCamelCase = max_size _lowerCamelCase = num_labels _lowerCamelCase = mask_feature_size def snake_case_ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.min_size, self.max_size] ).to( a__ ) _lowerCamelCase = torch.ones([self.batch_size, self.min_size, self.max_size] , device=a__ ) _lowerCamelCase = ( torch.rand([self.batch_size, self.num_labels, self.min_size, self.max_size] , device=a__ ) > 0.5 ).float() _lowerCamelCase = (torch.rand((self.batch_size, self.num_labels) , device=a__ ) > 0.5).long() _lowerCamelCase = self.get_config() return config, pixel_values, pixel_mask, mask_labels, class_labels def snake_case_ ( self ): return MaskFormerConfig.from_backbone_and_decoder_configs( backbone_config=SwinConfig( depths=[1, 1, 1, 1] , ) , decoder_config=DetrConfig( decoder_ffn_dim=1_28 , num_queries=self.num_queries , decoder_attention_heads=2 , d_model=self.mask_feature_size , ) , mask_feature_size=self.mask_feature_size , fpn_feature_size=self.mask_feature_size , num_channels=self.num_channels , num_labels=self.num_labels , ) def snake_case_ ( self ): _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase = {'pixel_values': pixel_values, 'pixel_mask': pixel_mask} return config, inputs_dict def snake_case_ ( self , a__ , a__ ): _lowerCamelCase = output.encoder_hidden_states _lowerCamelCase = output.pixel_decoder_hidden_states _lowerCamelCase = output.transformer_decoder_hidden_states self.parent.assertTrue(len(a__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(a__ ) , len(config.backbone_config.depths ) ) self.parent.assertTrue(len(a__ ) , config.decoder_config.decoder_layers ) def snake_case_ ( self , a__ , a__ , a__ , a__=False ): with torch.no_grad(): _lowerCamelCase = MaskFormerModel(config=a__ ) model.to(a__ ) model.eval() _lowerCamelCase = model(pixel_values=a__ , pixel_mask=a__ ) _lowerCamelCase = model(a__ , output_hidden_states=a__ ) # the correct shape of output.transformer_decoder_hidden_states ensure the correcteness of the # encoder and pixel decoder self.parent.assertEqual( output.transformer_decoder_last_hidden_state.shape , (self.batch_size, self.num_queries, self.mask_feature_size) , ) # let's ensure the other two hidden state exists self.parent.assertTrue(output.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(output.encoder_last_hidden_state is not None ) if output_hidden_states: self.check_output_hidden_state(a__ , a__ ) def snake_case_ ( self , a__ , a__ , a__ , a__ , a__ ): _lowerCamelCase = MaskFormerForInstanceSegmentation(config=a__ ) model.to(a__ ) model.eval() def comm_check_on_output(a__ ): # let's still check that all the required stuff is there self.parent.assertTrue(result.transformer_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.pixel_decoder_last_hidden_state is not None ) self.parent.assertTrue(result.encoder_last_hidden_state is not None ) # okay, now we need to check the logits shape # due to the encoder compression, masks have a //4 spatial size self.parent.assertEqual( result.masks_queries_logits.shape , (self.batch_size, self.num_queries, self.min_size // 4, self.max_size // 4) , ) # + 1 for null class self.parent.assertEqual( result.class_queries_logits.shape , (self.batch_size, self.num_queries, self.num_labels + 1) ) with torch.no_grad(): _lowerCamelCase = model(pixel_values=a__ , pixel_mask=a__ ) _lowerCamelCase = model(a__ ) comm_check_on_output(a__ ) _lowerCamelCase = model( pixel_values=a__ , pixel_mask=a__ , mask_labels=a__ , class_labels=a__ ) comm_check_on_output(a__ ) self.parent.assertTrue(result.loss is not None ) self.parent.assertEqual(result.loss.shape , torch.Size([1] ) ) @require_torch class __a ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): SCREAMING_SNAKE_CASE__ : Optional[int] = (MaskFormerModel, MaskFormerForInstanceSegmentation) if is_torch_available() else () SCREAMING_SNAKE_CASE__ : Optional[Any] = ( {"feature-extraction": MaskFormerModel, "image-segmentation": MaskFormerForInstanceSegmentation} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = False SCREAMING_SNAKE_CASE__ : Tuple = False SCREAMING_SNAKE_CASE__ : int = False SCREAMING_SNAKE_CASE__ : List[str] = False def snake_case_ ( self ): _lowerCamelCase = MaskFormerModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=a__ , has_text_modality=a__ ) def snake_case_ ( self ): self.config_tester.run_common_tests() def snake_case_ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(a__ , **a__ , output_hidden_states=a__ ) def snake_case_ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_maskformer_instance_segmentation_head_model(*a__ ) @unittest.skip(reason='MaskFormer does not use inputs_embeds' ) def snake_case_ ( self ): pass @unittest.skip(reason='MaskFormer does not have a get_input_embeddings method' ) def snake_case_ ( self ): pass @unittest.skip(reason='MaskFormer is not a generative model' ) def snake_case_ ( self ): pass @unittest.skip(reason='MaskFormer does not use token embeddings' ) def snake_case_ ( self ): pass @require_torch_multi_gpu @unittest.skip( reason='MaskFormer has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def snake_case_ ( self ): pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def snake_case_ ( self ): pass def snake_case_ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(a__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , a__ ) @slow def snake_case_ ( self ): for model_name in ["facebook/maskformer-swin-small-coco"]: _lowerCamelCase = MaskFormerModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) def snake_case_ ( self ): _lowerCamelCase = (self.model_tester.min_size,) * 2 _lowerCamelCase = { 'pixel_values': torch.randn((2, 3, *size) , device=a__ ), 'mask_labels': torch.randn((2, 10, *size) , device=a__ ), 'class_labels': torch.zeros(2 , 10 , device=a__ ).long(), } _lowerCamelCase = MaskFormerForInstanceSegmentation(MaskFormerConfig() ).to(a__ ) _lowerCamelCase = model(**a__ ) self.assertTrue(outputs.loss is not None ) def snake_case_ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.create_and_check_maskformer_model(a__ , **a__ , output_hidden_states=a__ ) def snake_case_ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(a__ ).to(a__ ) _lowerCamelCase = model(**a__ , output_attentions=a__ ) self.assertTrue(outputs.attentions is not None ) def snake_case_ ( self ): if not self.model_tester.is_training: return # only MaskFormerForInstanceSegmentation has the loss _lowerCamelCase = self.all_model_classes[1] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() _lowerCamelCase = model_class(a__ ) model.to(a__ ) model.train() _lowerCamelCase = model(a__ , mask_labels=a__ , class_labels=a__ ).loss loss.backward() def snake_case_ ( self ): # only MaskFormerForInstanceSegmentation has the loss _lowerCamelCase = self.all_model_classes[1] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs() _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = model_class(a__ ) model.to(a__ ) model.train() _lowerCamelCase = model(a__ , mask_labels=a__ , class_labels=a__ ) _lowerCamelCase = outputs.encoder_hidden_states[0] encoder_hidden_states.retain_grad() _lowerCamelCase = outputs.pixel_decoder_hidden_states[0] pixel_decoder_hidden_states.retain_grad() # we requires_grad=True in inputs_embeds (line 2152), the original implementation don't _lowerCamelCase = outputs.transformer_decoder_hidden_states[0] transformer_decoder_hidden_states.retain_grad() _lowerCamelCase = outputs.attentions[0] attentions.retain_grad() outputs.loss.backward(retain_graph=a__ ) self.assertIsNotNone(encoder_hidden_states.grad ) self.assertIsNotNone(pixel_decoder_hidden_states.grad ) self.assertIsNotNone(transformer_decoder_hidden_states.grad ) self.assertIsNotNone(attentions.grad ) A_ : Optional[int] =1E-4 def SCREAMING_SNAKE_CASE_ ( )-> Any: _lowerCamelCase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_vision @slow class __a ( unittest.TestCase ): @cached_property def snake_case_ ( self ): return ( MaskFormerImageProcessor.from_pretrained('facebook/maskformer-swin-small-coco' ) if is_vision_available() else None ) def snake_case_ ( self ): _lowerCamelCase = MaskFormerModel.from_pretrained('facebook/maskformer-swin-small-coco' ).to(a__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(a__ , return_tensors='pt' ).to(a__ ) _lowerCamelCase = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(a__ , (1, 3, 8_00, 10_88) ) with torch.no_grad(): _lowerCamelCase = model(**a__ ) _lowerCamelCase = torch.tensor( [[-0.0482, 0.9228, 0.4951], [-0.2547, 0.8017, 0.8527], [-0.0069, 0.3385, -0.0089]] ).to(a__ ) self.assertTrue( torch.allclose( outputs.encoder_last_hidden_state[0, 0, :3, :3] , a__ , atol=a__ ) ) _lowerCamelCase = torch.tensor( [[-0.8422, -0.8434, -0.9718], [-1.0144, -0.5565, -0.4195], [-1.0038, -0.4484, -0.1961]] ).to(a__ ) self.assertTrue( torch.allclose( outputs.pixel_decoder_last_hidden_state[0, 0, :3, :3] , a__ , atol=a__ ) ) _lowerCamelCase = torch.tensor( [[0.2852, -0.0159, 0.9735], [0.6254, 0.1858, 0.8529], [-0.0680, -0.4116, 1.8413]] ).to(a__ ) self.assertTrue( torch.allclose( outputs.transformer_decoder_last_hidden_state[0, :3, :3] , a__ , atol=a__ ) ) def snake_case_ ( self ): _lowerCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('facebook/maskformer-swin-small-coco' ) .to(a__ ) .eval() ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(a__ , return_tensors='pt' ).to(a__ ) _lowerCamelCase = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(a__ , (1, 3, 8_00, 10_88) ) with torch.no_grad(): _lowerCamelCase = model(**a__ ) # masks_queries_logits _lowerCamelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _lowerCamelCase = [ [-1.3737124, -1.7724937, -1.9364233], [-1.5977281, -1.9867939, -2.1523695], [-1.5795398, -1.9269832, -2.093942], ] _lowerCamelCase = torch.tensor(a__ ).to(a__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , a__ , atol=a__ ) ) # class_queries_logits _lowerCamelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _lowerCamelCase = torch.tensor( [ [1.6_512e00, -5.2_572e00, -3.3_519e00], [3.6_169e-02, -5.9_025e00, -2.9_313e00], [1.0_766e-04, -7.7_630e00, -5.1_263e00], ] ).to(a__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , a__ , atol=a__ ) ) def snake_case_ ( self ): _lowerCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('facebook/maskformer-resnet101-coco-stuff' ) .to(a__ ) .eval() ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(a__ , return_tensors='pt' ).to(a__ ) _lowerCamelCase = inputs['pixel_values'].shape # check size is divisible by 32 self.assertTrue((inputs_shape[-1] % 32) == 0 and (inputs_shape[-2] % 32) == 0 ) # check size self.assertEqual(a__ , (1, 3, 8_00, 10_88) ) with torch.no_grad(): _lowerCamelCase = model(**a__ ) # masks_queries_logits _lowerCamelCase = outputs.masks_queries_logits self.assertEqual( masks_queries_logits.shape , (1, model.config.decoder_config.num_queries, inputs_shape[-2] // 4, inputs_shape[-1] // 4) , ) _lowerCamelCase = [[-0.9046, -2.6366, -4.6062], [-3.4179, -5.7890, -8.8057], [-4.9179, -7.6560, -10.7711]] _lowerCamelCase = torch.tensor(a__ ).to(a__ ) self.assertTrue(torch.allclose(masks_queries_logits[0, 0, :3, :3] , a__ , atol=a__ ) ) # class_queries_logits _lowerCamelCase = outputs.class_queries_logits self.assertEqual( class_queries_logits.shape , (1, model.config.decoder_config.num_queries, model.config.num_labels + 1) ) _lowerCamelCase = torch.tensor( [[4.7188, -3.2585, -2.8857], [6.6871, -2.9181, -1.2487], [7.2449, -2.2764, -2.1874]] ).to(a__ ) self.assertTrue(torch.allclose(outputs.class_queries_logits[0, :3, :3] , a__ , atol=a__ ) ) def snake_case_ ( self ): _lowerCamelCase = ( MaskFormerForInstanceSegmentation.from_pretrained('facebook/maskformer-swin-small-coco' ) .to(a__ ) .eval() ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = image_processor( [np.zeros((3, 8_00, 13_33) ), np.zeros((3, 8_00, 13_33) )] , segmentation_maps=[np.zeros((3_84, 3_84) ).astype(np.floataa ), np.zeros((3_84, 3_84) ).astype(np.floataa )] , return_tensors='pt' , ) _lowerCamelCase = inputs['pixel_values'].to(a__ ) _lowerCamelCase = [el.to(a__ ) for el in inputs['mask_labels']] _lowerCamelCase = [el.to(a__ ) for el in inputs['class_labels']] with torch.no_grad(): _lowerCamelCase = model(**a__ ) self.assertTrue(outputs.loss is not None )
80
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ : int =logging.get_logger(__name__) A_ : Tuple ={ """ut/deta""": """https://huggingface.co/ut/deta/resolve/main/config.json""", } class __a ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE__ : int = "deta" SCREAMING_SNAKE_CASE__ : Union[str, Any] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self , a__=None , a__=9_00 , a__=20_48 , a__=6 , a__=20_48 , a__=8 , a__=6 , a__=10_24 , a__=8 , a__=0.0 , a__=True , a__="relu" , a__=2_56 , a__=0.1 , a__=0.0 , a__=0.0 , a__=0.02 , a__=1.0 , a__=True , a__=False , a__="sine" , a__=5 , a__=4 , a__=4 , a__=True , a__=3_00 , a__=True , a__=True , a__=1 , a__=5 , a__=2 , a__=1 , a__=1 , a__=5 , a__=2 , a__=0.1 , a__=0.25 , **a__ , ): if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _lowerCamelCase = CONFIG_MAPPING['resnet'](out_features=['stage2', 'stage3', 'stage4'] ) else: if isinstance(a__ , a__ ): _lowerCamelCase = backbone_config.pop('model_type' ) _lowerCamelCase = CONFIG_MAPPING[backbone_model_type] _lowerCamelCase = config_class.from_dict(a__ ) _lowerCamelCase = backbone_config _lowerCamelCase = num_queries _lowerCamelCase = max_position_embeddings _lowerCamelCase = d_model _lowerCamelCase = encoder_ffn_dim _lowerCamelCase = encoder_layers _lowerCamelCase = encoder_attention_heads _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_attention_heads _lowerCamelCase = dropout _lowerCamelCase = attention_dropout _lowerCamelCase = activation_dropout _lowerCamelCase = activation_function _lowerCamelCase = init_std _lowerCamelCase = init_xavier_std _lowerCamelCase = encoder_layerdrop _lowerCamelCase = auxiliary_loss _lowerCamelCase = position_embedding_type # deformable attributes _lowerCamelCase = num_feature_levels _lowerCamelCase = encoder_n_points _lowerCamelCase = decoder_n_points _lowerCamelCase = two_stage _lowerCamelCase = two_stage_num_proposals _lowerCamelCase = with_box_refine _lowerCamelCase = assign_first_stage if two_stage is True and with_box_refine is False: raise ValueError('If two_stage is True, with_box_refine must be True.' ) # Hungarian matcher _lowerCamelCase = class_cost _lowerCamelCase = bbox_cost _lowerCamelCase = giou_cost # Loss coefficients _lowerCamelCase = mask_loss_coefficient _lowerCamelCase = dice_loss_coefficient _lowerCamelCase = bbox_loss_coefficient _lowerCamelCase = giou_loss_coefficient _lowerCamelCase = eos_coefficient _lowerCamelCase = focal_alpha super().__init__(is_encoder_decoder=a__ , **a__ ) @property def snake_case_ ( self ): return self.encoder_attention_heads @property def snake_case_ ( self ): return self.d_model def snake_case_ ( self ): _lowerCamelCase = copy.deepcopy(self.__dict__ ) _lowerCamelCase = self.backbone_config.to_dict() _lowerCamelCase = self.__class__.model_type return output
80
1
"""simple docstring""" import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print('''Googling.....''') _SCREAMING_SNAKE_CASE : Dict = '''https://www.google.com/search?q=''' + ''' '''.join(sys.argv[1:]) _SCREAMING_SNAKE_CASE : str = requests.get(url, headers={'''UserAgent''': UserAgent().random}) # res.raise_for_status() with open('''project1a.html''', '''wb''') as out_file: # only for knowing the class for data in res.iter_content(1_0000): out_file.write(data) _SCREAMING_SNAKE_CASE : Union[str, Any] = BeautifulSoup(res.text, '''html.parser''') _SCREAMING_SNAKE_CASE : Union[str, Any] = list(soup.select('''.eZt8xd'''))[:5] print(len(links)) for link in links: if link.text == "Maps": webbrowser.open(link.get('''href''')) else: webbrowser.open(F'''https://google.com{link.get('href')}''')
183
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class a ( unittest.TestCase ): def __init__( self : Any , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : List[str]=7 , __SCREAMING_SNAKE_CASE : Dict=3 , __SCREAMING_SNAKE_CASE : Dict=18 , __SCREAMING_SNAKE_CASE : Union[str, Any]=30 , __SCREAMING_SNAKE_CASE : Optional[Any]=400 , __SCREAMING_SNAKE_CASE : Dict=True , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : Any=True , ) -> str: lowerCamelCase_ = size if size is not None else {'height': 18, 'width': 18} lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = image_size lowerCamelCase_ = min_resolution lowerCamelCase_ = max_resolution lowerCamelCase_ = do_resize lowerCamelCase_ = size lowerCamelCase_ = apply_ocr def UpperCamelCase ( self : int ) -> Tuple: return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class a ( __snake_case , unittest.TestCase ): SCREAMING_SNAKE_CASE : str = LayoutLMvaImageProcessor if is_pytesseract_available() else None def UpperCamelCase ( self : List[str] ) -> int: lowerCamelCase_ = LayoutLMvaImageProcessingTester(self ) @property def UpperCamelCase ( self : Optional[Any] ) -> Any: return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase ( self : Tuple ) -> Optional[Any]: lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , 'do_resize' ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , 'size' ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , 'apply_ocr' ) ) def UpperCamelCase ( self : Any ) -> Any: lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 18, 'width': 18} ) lowerCamelCase_ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'height': 42, 'width': 42} ) def UpperCamelCase ( self : Dict ) -> Any: pass def UpperCamelCase ( self : int ) -> Dict: # Initialize image_processing lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , Image.Image ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ) self.assertEqual( encoding.pixel_values.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) self.assertIsInstance(encoding.words , __SCREAMING_SNAKE_CASE ) self.assertIsInstance(encoding.boxes , __SCREAMING_SNAKE_CASE ) # Test batched lowerCamelCase_ = image_processing(__SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def UpperCamelCase ( self : Optional[int] ) -> Optional[Any]: # Initialize image_processing lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__SCREAMING_SNAKE_CASE , numpify=__SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , np.ndarray ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) # Test batched lowerCamelCase_ = image_processing(__SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def UpperCamelCase ( self : Dict ) -> int: # Initialize image_processing lowerCamelCase_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCamelCase_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__SCREAMING_SNAKE_CASE , torchify=__SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , torch.Tensor ) # Test not batched input lowerCamelCase_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) # Test batched lowerCamelCase_ = image_processing(__SCREAMING_SNAKE_CASE , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['height'], self.image_processor_tester.size['width'], ) , ) def UpperCamelCase ( self : Dict ) -> Any: # with apply_OCR = True lowerCamelCase_ = LayoutLMvaImageProcessor() from datasets import load_dataset lowerCamelCase_ = load_dataset('hf-internal-testing/fixtures_docvqa' , split='test' ) lowerCamelCase_ = Image.open(ds[0]['file'] ).convert('RGB' ) lowerCamelCase_ = image_processing(__SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) ) self.assertEqual(len(encoding.words ) , len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 lowerCamelCase_ = [['11:14', 'to', '11:39', 'a.m', '11:39', 'to', '11:44', 'a.m.', '11:44', 'a.m.', 'to', '12:25', 'p.m.', '12:25', 'to', '12:58', 'p.m.', '12:58', 'to', '4:00', 'p.m.', '2:00', 'to', '5:00', 'p.m.', 'Coffee', 'Break', 'Coffee', 'will', 'be', 'served', 'for', 'men', 'and', 'women', 'in', 'the', 'lobby', 'adjacent', 'to', 'exhibit', 'area.', 'Please', 'move', 'into', 'exhibit', 'area.', '(Exhibits', 'Open)', 'TRRF', 'GENERAL', 'SESSION', '(PART', '|)', 'Presiding:', 'Lee', 'A.', 'Waller', 'TRRF', 'Vice', 'President', '“Introductory', 'Remarks”', 'Lee', 'A.', 'Waller,', 'TRRF', 'Vice', 'Presi-', 'dent', 'Individual', 'Interviews', 'with', 'TRRF', 'Public', 'Board', 'Members', 'and', 'Sci-', 'entific', 'Advisory', 'Council', 'Mem-', 'bers', 'Conducted', 'by', 'TRRF', 'Treasurer', 'Philip', 'G.', 'Kuehn', 'to', 'get', 'answers', 'which', 'the', 'public', 'refrigerated', 'warehousing', 'industry', 'is', 'looking', 'for.', 'Plus', 'questions', 'from', 'the', 'floor.', 'Dr.', 'Emil', 'M.', 'Mrak,', 'University', 'of', 'Cal-', 'ifornia,', 'Chairman,', 'TRRF', 'Board;', 'Sam', 'R.', 'Cecil,', 'University', 'of', 'Georgia', 'College', 'of', 'Agriculture;', 'Dr.', 'Stanley', 'Charm,', 'Tufts', 'University', 'School', 'of', 'Medicine;', 'Dr.', 'Robert', 'H.', 'Cotton,', 'ITT', 'Continental', 'Baking', 'Company;', 'Dr.', 'Owen', 'Fennema,', 'University', 'of', 'Wis-', 'consin;', 'Dr.', 'Robert', 'E.', 'Hardenburg,', 'USDA.', 'Questions', 'and', 'Answers', 'Exhibits', 'Open', 'Capt.', 'Jack', 'Stoney', 'Room', 'TRRF', 'Scientific', 'Advisory', 'Council', 'Meeting', 'Ballroom', 'Foyer']] # noqa: E231 lowerCamelCase_ = [[[141, 57, 214, 69], [228, 58, 252, 69], [141, 75, 216, 88], [230, 79, 280, 88], [142, 260, 218, 273], [230, 261, 255, 273], [143, 279, 218, 290], [231, 282, 290, 291], [143, 342, 218, 354], [231, 345, 289, 355], [202, 362, 227, 373], [143, 379, 220, 392], [231, 382, 291, 394], [144, 714, 220, 726], [231, 715, 256, 726], [144, 732, 220, 745], [232, 736, 291, 747], [144, 769, 218, 782], [231, 770, 256, 782], [141, 788, 202, 801], [215, 791, 274, 804], [143, 826, 204, 838], [215, 826, 240, 838], [142, 844, 202, 857], [215, 847, 274, 859], [334, 57, 427, 69], [440, 57, 522, 69], [369, 75, 461, 88], [469, 75, 516, 88], [528, 76, 562, 88], [570, 76, 667, 88], [675, 75, 711, 87], [721, 79, 778, 88], [789, 75, 840, 88], [369, 97, 470, 107], [484, 94, 507, 106], [518, 94, 562, 107], [576, 94, 655, 110], [668, 94, 792, 109], [804, 95, 829, 107], [369, 113, 465, 125], [477, 116, 547, 125], [562, 113, 658, 125], [671, 116, 748, 125], [761, 113, 811, 125], [369, 131, 465, 143], [477, 133, 548, 143], [563, 130, 698, 145], [710, 130, 802, 146], [336, 171, 412, 183], [423, 171, 572, 183], [582, 170, 716, 184], [728, 171, 817, 187], [829, 171, 844, 186], [338, 197, 482, 212], [507, 196, 557, 209], [569, 196, 595, 208], [610, 196, 702, 209], [505, 214, 583, 226], [595, 214, 656, 227], [670, 215, 807, 227], [335, 259, 543, 274], [556, 259, 708, 272], [372, 279, 422, 291], [435, 279, 460, 291], [474, 279, 574, 292], [587, 278, 664, 291], [676, 278, 738, 291], [751, 279, 834, 291], [372, 298, 434, 310], [335, 341, 483, 354], [497, 341, 655, 354], [667, 341, 728, 354], [740, 341, 825, 354], [335, 360, 430, 372], [442, 360, 534, 372], [545, 359, 687, 372], [697, 360, 754, 372], [765, 360, 823, 373], [334, 378, 428, 391], [440, 378, 577, 394], [590, 378, 705, 391], [720, 378, 801, 391], [334, 397, 400, 409], [370, 416, 529, 429], [544, 416, 576, 432], [587, 416, 665, 428], [677, 416, 814, 429], [372, 435, 452, 450], [465, 434, 495, 447], [511, 434, 600, 447], [611, 436, 637, 447], [649, 436, 694, 451], [705, 438, 824, 447], [369, 453, 452, 466], [464, 454, 509, 466], [522, 453, 611, 469], [625, 453, 792, 469], [370, 472, 556, 488], [570, 472, 684, 487], [697, 472, 718, 485], [732, 472, 835, 488], [369, 490, 411, 503], [425, 490, 484, 503], [496, 490, 635, 506], [645, 490, 707, 503], [718, 491, 761, 503], [771, 490, 840, 503], [336, 510, 374, 521], [388, 510, 447, 522], [460, 510, 489, 521], [503, 510, 580, 522], [592, 509, 736, 525], [745, 509, 770, 522], [781, 509, 840, 522], [338, 528, 434, 541], [448, 528, 596, 541], [609, 527, 687, 540], [700, 528, 792, 541], [336, 546, 397, 559], [407, 546, 431, 559], [443, 546, 525, 560], [537, 546, 680, 562], [688, 546, 714, 559], [722, 546, 837, 562], [336, 565, 449, 581], [461, 565, 485, 577], [497, 565, 665, 581], [681, 565, 718, 577], [732, 565, 837, 580], [337, 584, 438, 597], [452, 583, 521, 596], [535, 584, 677, 599], [690, 583, 787, 596], [801, 583, 825, 596], [338, 602, 478, 615], [492, 602, 530, 614], [543, 602, 638, 615], [650, 602, 676, 614], [688, 602, 788, 615], [802, 602, 843, 614], [337, 621, 502, 633], [516, 621, 615, 637], [629, 621, 774, 636], [789, 621, 827, 633], [337, 639, 418, 652], [432, 640, 571, 653], [587, 639, 731, 655], [743, 639, 769, 652], [780, 639, 841, 652], [338, 658, 440, 673], [455, 658, 491, 670], [508, 658, 602, 671], [616, 658, 638, 670], [654, 658, 835, 674], [337, 677, 429, 689], [337, 714, 482, 726], [495, 714, 548, 726], [561, 714, 683, 726], [338, 770, 461, 782], [474, 769, 554, 785], [489, 788, 562, 803], [576, 788, 643, 801], [656, 787, 751, 804], [764, 788, 844, 801], [334, 825, 421, 838], [430, 824, 574, 838], [584, 824, 723, 841], [335, 844, 450, 857], [464, 843, 583, 860], [628, 862, 755, 875], [769, 861, 848, 878]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words , __SCREAMING_SNAKE_CASE ) self.assertListEqual(encoding.boxes , __SCREAMING_SNAKE_CASE ) # with apply_OCR = False lowerCamelCase_ = LayoutLMvaImageProcessor(apply_ocr=__SCREAMING_SNAKE_CASE ) lowerCamelCase_ = image_processing(__SCREAMING_SNAKE_CASE , return_tensors='pt' ) self.assertEqual(encoding.pixel_values.shape , (1, 3, 224, 224) )
183
1
'''simple docstring''' import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __a = Mapping[str, np.ndarray] __a = Mapping[str, Any] # Is a nested dict. __a = 0.01 @dataclasses.dataclass(frozen=_a ) class UpperCAmelCase_ : """simple docstring""" lowercase = 42 # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. lowercase = 42 # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. lowercase = 42 # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. lowercase = 42 # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. lowercase = 42 # [num_res, num_atom_type] # Chain indices for multi-chain predictions lowercase = None # Optional remark about the protein. Included as a comment in output PDB # files lowercase = None # Templates used to generate this protein (prediction-only) lowercase = None # Chain corresponding to each parent lowercase = None def __snake_case( _lowerCAmelCase ) -> Protein: snake_case__ : Optional[Any] = r"""(\[[A-Z]+\]\n)""" snake_case__ : List[str] = [tag.strip() for tag in re.split(_lowerCAmelCase , _lowerCAmelCase ) if len(_lowerCAmelCase ) > 0] snake_case__ : Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split("""\n""" ) for l in tags[1::2]] ) snake_case__ : List[str] = ["N", "CA", "C"] snake_case__ : Optional[Any] = None snake_case__ : List[str] = None snake_case__ : Tuple = None for g in groups: if "[PRIMARY]" == g[0]: snake_case__ : Optional[Any] = g[1][0].strip() for i in range(len(_lowerCAmelCase ) ): if seq[i] not in residue_constants.restypes: snake_case__ : Optional[int] = """X""" # FIXME: strings are immutable snake_case__ : str = np.array( [residue_constants.restype_order.get(_lowerCAmelCase , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: snake_case__ : List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(_lowerCAmelCase , g[1][axis].split() ) ) ) snake_case__ : Union[str, Any] = np.array(_lowerCAmelCase ) snake_case__ : str = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(_lowerCAmelCase ): snake_case__ : List[Any] = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: snake_case__ : Optional[int] = np.array(list(map({"""-""": 0, """+""": 1}.get , g[1][0].strip() ) ) ) snake_case__ : List[Any] = np.zeros( ( len(_lowerCAmelCase ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(_lowerCAmelCase ): snake_case__ : Union[str, Any] = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=_lowerCAmelCase , atom_mask=_lowerCAmelCase , aatype=_lowerCAmelCase , residue_index=np.arange(len(_lowerCAmelCase ) ) , b_factors=_lowerCAmelCase , ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase = 0 ) -> List[str]: snake_case__ : List[str] = [] snake_case__ : Dict = prot.remark if remark is not None: pdb_headers.append(f"REMARK {remark}" ) snake_case__ : Optional[int] = prot.parents snake_case__ : int = prot.parents_chain_index if parents is not None and parents_chain_index is not None: snake_case__ : Any = [p for i, p in zip(_lowerCAmelCase , _lowerCAmelCase ) if i == chain_id] if parents is None or len(_lowerCAmelCase ) == 0: snake_case__ : str = ["""N/A"""] pdb_headers.append(f"PARENT {' '.join(_lowerCAmelCase )}" ) return pdb_headers def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> str: snake_case__ : List[str] = [] snake_case__ : Union[str, Any] = pdb_str.split("""\n""" ) snake_case__ : Optional[Any] = prot.remark if remark is not None: out_pdb_lines.append(f"REMARK {remark}" ) snake_case__ : List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: snake_case__ : Any = [] if prot.parents_chain_index is not None: snake_case__ : Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(_lowerCAmelCase ) , [] ) parent_dict[str(_lowerCAmelCase )].append(_lowerCAmelCase ) snake_case__ : Tuple = max([int(_lowerCAmelCase ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): snake_case__ : Optional[int] = parent_dict.get(str(_lowerCAmelCase ) , ["""N/A"""] ) parents_per_chain.append(_lowerCAmelCase ) else: parents_per_chain.append(list(prot.parents ) ) else: snake_case__ : str = [["""N/A"""]] def make_parent_line(_lowerCAmelCase ) -> str: return f"PARENT {' '.join(_lowerCAmelCase )}" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) snake_case__ : List[str] = 0 for i, l in enumerate(_lowerCAmelCase ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(_lowerCAmelCase ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(_lowerCAmelCase ): snake_case__ : Tuple = parents_per_chain[chain_counter] else: snake_case__ : int = ["""N/A"""] out_pdb_lines.append(make_parent_line(_lowerCAmelCase ) ) return "\n".join(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> str: snake_case__ : Optional[int] = residue_constants.restypes + ["""X"""] def res_atoa(_lowerCAmelCase ) -> str: return residue_constants.restype_atoa.get(restypes[r] , """UNK""" ) snake_case__ : int = residue_constants.atom_types snake_case__ : List[str] = [] snake_case__ : Any = prot.atom_mask snake_case__ : str = prot.aatype snake_case__ : Any = prot.atom_positions snake_case__ : List[Any] = prot.residue_index.astype(np.intaa ) snake_case__ : Union[str, Any] = prot.b_factors snake_case__ : Dict = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError("""Invalid aatypes.""" ) snake_case__ : List[Any] = get_pdb_headers(_lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: pdb_lines.extend(_lowerCAmelCase ) snake_case__ : Tuple = aatype.shape[0] snake_case__ : List[str] = 1 snake_case__ : List[Any] = 0 snake_case__ : int = string.ascii_uppercase snake_case__ : Union[str, Any] = None # Add all atom sites. for i in range(_lowerCAmelCase ): snake_case__ : Optional[int] = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(_lowerCAmelCase , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue snake_case__ : Dict = """ATOM""" snake_case__ : List[str] = atom_name if len(_lowerCAmelCase ) == 4 else f" {atom_name}" snake_case__ : Dict = """""" snake_case__ : List[Any] = """""" snake_case__ : Union[str, Any] = 1.00 snake_case__ : Union[str, Any] = atom_name[0] # Protein supports only C, N, O, S, this works. snake_case__ : Optional[int] = """""" snake_case__ : List[Any] = """A""" if chain_index is not None: snake_case__ : Dict = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! snake_case__ : Optional[Any] = ( f"{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}" f"{res_name_a:>3} {chain_tag:>1}" f"{residue_index[i]:>4}{insertion_code:>1} " f"{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}" f"{occupancy:>6.2f}{b_factor:>6.2f} " f"{element:>2}{charge:>2}" ) pdb_lines.append(_lowerCAmelCase ) atom_index += 1 snake_case__ : List[str] = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: snake_case__ : Any = True snake_case__ : List[str] = chain_index[i + 1] if should_terminate: # Close the chain. snake_case__ : int = """TER""" snake_case__ : Union[str, Any] = ( f"{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}" ) pdb_lines.append(_lowerCAmelCase ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(_lowerCAmelCase , _lowerCAmelCase ) ) pdb_lines.append("""END""" ) pdb_lines.append("""""" ) return "\n".join(_lowerCAmelCase ) def __snake_case( _lowerCAmelCase ) -> np.ndarray: return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def __snake_case( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , _lowerCAmelCase = None , ) -> Protein: return Protein( aatype=features["""aatype"""] , atom_positions=result["""final_atom_positions"""] , atom_mask=result["""final_atom_mask"""] , residue_index=features["""residue_index"""] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result["""final_atom_mask"""] ) , chain_index=_lowerCAmelCase , remark=_lowerCAmelCase , parents=_lowerCAmelCase , parents_chain_index=_lowerCAmelCase , )
43
'''simple docstring''' def __snake_case( _lowerCAmelCase ) -> int: if not isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError("""Input value must be an 'int' type""" ) snake_case__ : List[str] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
43
1
"""simple docstring""" import heapq as hq import math from collections.abc import Iterator class _UpperCamelCase : '''simple docstring''' def __init__( self , __a ): __lowerCAmelCase = str(id_ ) __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = [] __lowerCAmelCase = {} # {vertex:distance} def __lt__( self , __a ): return self.key < other.key def __repr__( self ): return self.id def snake_case ( self , __a ): self.neighbors.append(__a ) def snake_case ( self , __a , __a ): __lowerCAmelCase = weight def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' graph[a - 1].add_neighbor(graph[b - 1] ) graph[b - 1].add_neighbor(graph[a - 1] ) # add the edges: graph[a - 1].add_edge(graph[b - 1] , _UpperCamelCase ) graph[b - 1].add_edge(graph[a - 1] , _UpperCamelCase ) def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = [] for u in graph: __lowerCAmelCase = math.inf __lowerCAmelCase = None __lowerCAmelCase = 0 __lowerCAmelCase = graph[:] while q: __lowerCAmelCase = min(_UpperCamelCase ) q.remove(_UpperCamelCase ) for v in u.neighbors: if (v in q) and (u.edges[v.id] < v.key): __lowerCAmelCase = u __lowerCAmelCase = u.edges[v.id] for i in range(1 , len(_UpperCamelCase ) ): a.append((int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) ) return a def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' for u in graph: __lowerCAmelCase = math.inf __lowerCAmelCase = None __lowerCAmelCase = 0 __lowerCAmelCase = list(_UpperCamelCase ) hq.heapify(_UpperCamelCase ) while h: __lowerCAmelCase = hq.heappop(_UpperCamelCase ) for v in u.neighbors: if (v in h) and (u.edges[v.id] < v.key): __lowerCAmelCase = u __lowerCAmelCase = u.edges[v.id] hq.heapify(_UpperCamelCase ) for i in range(1 , len(_UpperCamelCase ) ): yield (int(graph[i].id ) + 1, int(graph[i].pi.id ) + 1) def _lowerCamelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
57
'''simple docstring''' import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 lowercase : str = get_tests_dir('fixtures/dummy_feature_extractor_config.json') lowercase : str = get_tests_dir('fixtures/vocab.json') lowercase : int = get_tests_dir('fixtures') class A ( unittest.TestCase ): __magic_name__ = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''bla''', '''blou'''] def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" A : Tuple = 0 def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : List[Any] = AutoProcessor.from_pretrained('''facebook/wav2vec2-base-960h''' ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: A : Union[str, Any] = WavaVecaConfig() A : List[str] = AutoProcessor.from_pretrained('''facebook/wav2vec2-base-960h''' ) # save in new folder model_config.save_pretrained(SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) A : List[str] = AutoProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> int: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(SCREAMING_SNAKE_CASE , os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) copyfile(SCREAMING_SNAKE_CASE , os.path.join(SCREAMING_SNAKE_CASE , '''vocab.json''' ) ) A : Optional[Any] = AutoProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: A : Dict = WavaVecaFeatureExtractor() A : List[str] = AutoTokenizer.from_pretrained('''facebook/wav2vec2-base-960h''' ) A : str = WavaVecaProcessor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # save in new folder processor.save_pretrained(SCREAMING_SNAKE_CASE ) # drop `processor_class` in tokenizer with open(os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , '''r''' ) as f: A : Dict = json.load(SCREAMING_SNAKE_CASE ) config_dict.pop('''processor_class''' ) with open(os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , '''w''' ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE ) ) A : Optional[Any] = AutoProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> int: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: A : List[Any] = WavaVecaFeatureExtractor() A : List[Any] = AutoTokenizer.from_pretrained('''facebook/wav2vec2-base-960h''' ) A : str = WavaVecaProcessor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # save in new folder processor.save_pretrained(SCREAMING_SNAKE_CASE ) # drop `processor_class` in feature extractor with open(os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , '''r''' ) as f: A : str = json.load(SCREAMING_SNAKE_CASE ) config_dict.pop('''processor_class''' ) with open(os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , '''w''' ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE ) ) A : str = AutoProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: A : str = WavaVecaConfig(processor_class='''Wav2Vec2Processor''' ) model_config.save_pretrained(SCREAMING_SNAKE_CASE ) # copy relevant files copyfile(SCREAMING_SNAKE_CASE , os.path.join(SCREAMING_SNAKE_CASE , '''vocab.json''' ) ) # create emtpy sample processor with open(os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , '''w''' ) as f: f.write('''{}''' ) A : List[str] = AutoProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" with self.assertRaises(SCREAMING_SNAKE_CASE ): A : Optional[Any] = AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(SCREAMING_SNAKE_CASE ): A : Union[str, Any] = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=SCREAMING_SNAKE_CASE ) A : Union[str, Any] = AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=SCREAMING_SNAKE_CASE ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) A : List[str] = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , '''NewFeatureExtractor''' ) A : Tuple = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) # Test we can also load the slow version A : List[str] = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=SCREAMING_SNAKE_CASE , use_fast=SCREAMING_SNAKE_CASE ) A : int = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , '''NewTokenizer''' ) else: self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" try: AutoConfig.register('''custom''' , SCREAMING_SNAKE_CASE ) AutoFeatureExtractor.register(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) AutoTokenizer.register(SCREAMING_SNAKE_CASE , slow_tokenizer_class=SCREAMING_SNAKE_CASE ) AutoProcessor.register(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(SCREAMING_SNAKE_CASE ): AutoProcessor.register(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Now that the config is registered, it can be used as any other config with the auto-API A : List[Any] = CustomFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dir: A : Tuple = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.txt''' ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) A : Optional[int] = CustomTokenizer(SCREAMING_SNAKE_CASE ) A : Any = CustomProcessor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(SCREAMING_SNAKE_CASE ) A : List[str] = AutoProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) self.assertIsInstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" class A ( __snake_case ): __magic_name__ = False class A ( __snake_case ): __magic_name__ = False class A ( __snake_case ): __magic_name__ = '''AutoFeatureExtractor''' __magic_name__ = '''AutoTokenizer''' __magic_name__ = False try: AutoConfig.register('''custom''' , SCREAMING_SNAKE_CASE ) AutoFeatureExtractor.register(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) AutoTokenizer.register(SCREAMING_SNAKE_CASE , slow_tokenizer_class=SCREAMING_SNAKE_CASE ) AutoProcessor.register(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # If remote code is not set, the default is to use local classes. A : Union[str, Any] = AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. A : Optional[int] = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=SCREAMING_SNAKE_CASE ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. A : Tuple = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=SCREAMING_SNAKE_CASE ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : int = AutoProcessor.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) self.assertEqual(processor.__class__.__name__ , '''BertTokenizerFast''' ) def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : Optional[int] = AutoProcessor.from_pretrained('''hf-internal-testing/tiny-random-convnext''' ) self.assertEqual(processor.__class__.__name__ , '''ConvNextImageProcessor''' ) @is_staging_test class A ( unittest.TestCase ): __magic_name__ = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''bla''', '''blou'''] @classmethod def __lowerCAmelCase ( cls ) -> Dict: """simple docstring""" A : Optional[int] = TOKEN HfFolder.save_token(SCREAMING_SNAKE_CASE ) @classmethod def __lowerCAmelCase ( cls ) -> Any: """simple docstring""" try: delete_repo(token=cls._token , repo_id='''test-processor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-processor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-processor''' ) except HTTPError: pass def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : Union[str, Any] = WavaVecaProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(SCREAMING_SNAKE_CASE , '''test-processor''' ) , push_to_hub=SCREAMING_SNAKE_CASE , use_auth_token=self._token ) A : int = WavaVecaProcessor.from_pretrained(F'{USER}/test-processor' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(SCREAMING_SNAKE_CASE , getattr(new_processor.feature_extractor , SCREAMING_SNAKE_CASE ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : Tuple = WavaVecaProcessor.from_pretrained(SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(SCREAMING_SNAKE_CASE , '''test-processor-org''' ) , push_to_hub=SCREAMING_SNAKE_CASE , use_auth_token=self._token , organization='''valid_org''' , ) A : int = WavaVecaProcessor.from_pretrained('''valid_org/test-processor-org''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(SCREAMING_SNAKE_CASE , getattr(new_processor.feature_extractor , SCREAMING_SNAKE_CASE ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def __lowerCAmelCase ( self ) -> str: """simple docstring""" CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() A : Any = CustomFeatureExtractor.from_pretrained(SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dir: A : Union[str, Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.txt''' ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) A : str = CustomTokenizer(SCREAMING_SNAKE_CASE ) A : Union[str, Any] = CustomProcessor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(F'{USER}/test-dynamic-processor' , token=self._token ) A : List[str] = Repository(SCREAMING_SNAKE_CASE , clone_from=F'{USER}/test-dynamic-processor' , token=self._token ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { '''AutoFeatureExtractor''': '''custom_feature_extraction.CustomFeatureExtractor''', '''AutoProcessor''': '''custom_processing.CustomProcessor''', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(SCREAMING_SNAKE_CASE , '''tokenizer_config.json''' ) ) as f: A : Dict = json.load(SCREAMING_SNAKE_CASE ) self.assertDictEqual( tokenizer_config['''auto_map'''] , { '''AutoTokenizer''': ['''custom_tokenization.CustomTokenizer''', None], '''AutoProcessor''': '''custom_processing.CustomProcessor''', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(SCREAMING_SNAKE_CASE , '''custom_feature_extraction.py''' ) ) ) self.assertTrue(os.path.isfile(os.path.join(SCREAMING_SNAKE_CASE , '''custom_tokenization.py''' ) ) ) self.assertTrue(os.path.isfile(os.path.join(SCREAMING_SNAKE_CASE , '''custom_processing.py''' ) ) ) repo.push_to_hub() A : Optional[int] = AutoProcessor.from_pretrained(F'{USER}/test-dynamic-processor' , trust_remote_code=SCREAMING_SNAKE_CASE ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , '''CustomProcessor''' )
3
0
def SCREAMING_SNAKE_CASE__ ( lowercase ) -> list[int]: snake_case : Optional[int] = [0 for i in range(len(lowercase ) )] # initialize interval's left pointer and right pointer snake_case , snake_case : Dict = 0, 0 for i in range(1 ,len(lowercase ) ): # case when current index is inside the interval if i <= right_pointer: snake_case : Dict = min(right_pointer - i + 1 ,z_result[i - left_pointer] ) snake_case : Optional[Any] = min_edge while go_next(lowercase ,lowercase ,lowercase ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: snake_case , snake_case : str = i, i + z_result[i] - 1 return z_result def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ) -> bool: return i + z_result[i] < len(lowercase ) and s[z_result[i]] == s[i + z_result[i]] def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> int: snake_case : List[str] = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string snake_case : Any = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(lowercase ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
354
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase : str = { 'configuration_roformer': ['ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoFormerConfig', 'RoFormerOnnxConfig'], 'tokenization_roformer': ['RoFormerTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[Any] = ['RoFormerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : Dict = [ 'ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'RoFormerForCausalLM', 'RoFormerForMaskedLM', 'RoFormerForMultipleChoice', 'RoFormerForQuestionAnswering', 'RoFormerForSequenceClassification', 'RoFormerForTokenClassification', 'RoFormerLayer', 'RoFormerModel', 'RoFormerPreTrainedModel', 'load_tf_weights_in_roformer', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : Any = [ 'TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRoFormerForCausalLM', 'TFRoFormerForMaskedLM', 'TFRoFormerForMultipleChoice', 'TFRoFormerForQuestionAnswering', 'TFRoFormerForSequenceClassification', 'TFRoFormerForTokenClassification', 'TFRoFormerLayer', 'TFRoFormerModel', 'TFRoFormerPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : Union[str, Any] = [ 'FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxRoFormerForMaskedLM', 'FlaxRoFormerForMultipleChoice', 'FlaxRoFormerForQuestionAnswering', 'FlaxRoFormerForSequenceClassification', 'FlaxRoFormerForTokenClassification', 'FlaxRoFormerModel', 'FlaxRoFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys lowerCamelCase : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
176
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : str = logging.get_logger(__name__) __lowerCAmelCase : Tuple = { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/config.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/config.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/config.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/config.json', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json', 'roberta-large-openai-detector': 'https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json', } class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = """roberta""" def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any]=5_02_65 , __lowerCamelCase : Optional[Any]=7_68 , __lowerCamelCase : List[Any]=12 , __lowerCamelCase : Union[str, Any]=12 , __lowerCamelCase : int=30_72 , __lowerCamelCase : Any="gelu" , __lowerCamelCase : int=0.1 , __lowerCamelCase : Optional[Any]=0.1 , __lowerCamelCase : int=5_12 , __lowerCamelCase : Tuple=2 , __lowerCamelCase : Tuple=0.02 , __lowerCamelCase : Union[str, Any]=1e-12 , __lowerCamelCase : Any=1 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Optional[int]="absolute" , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : int=None , **__lowerCamelCase : List[str] , ) -> Tuple: super().__init__(pad_token_id=__lowerCamelCase , bos_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , **__lowerCamelCase ) a = vocab_size a = hidden_size a = num_hidden_layers a = num_attention_heads a = hidden_act a = intermediate_size a = hidden_dropout_prob a = attention_probs_dropout_prob a = max_position_embeddings a = type_vocab_size a = initializer_range a = layer_norm_eps a = position_embedding_type a = use_cache a = classifier_dropout class snake_case__ (_UpperCamelCase ): """simple docstring""" @property def __UpperCAmelCase ( self : List[str] ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": a = {0: "batch", 1: "choice", 2: "sequence"} else: a = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
107
from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
107
1
"""simple docstring""" import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC _A = parse(importlib.metadata.version("""torch""")) def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Union[str, Any]: if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) UpperCAmelCase__ : int = STR_OPERATION_TO_FUNC[operation] if isinstance(lowerCAmelCase , lowerCAmelCase ): UpperCAmelCase__ : Union[str, Any] = parse(importlib.metadata.version(lowerCAmelCase ) ) return operation(lowerCAmelCase , parse(lowerCAmelCase ) ) def a__ ( lowerCAmelCase , lowerCAmelCase ) -> Optional[Any]: return compare_versions(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase )
166
"""simple docstring""" import sys import turtle def a__ ( lowerCAmelCase , lowerCAmelCase ) -> tuple[float, float]: return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ) -> None: my_pen.up() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.down() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) if depth == 0: return triangle(lowerCAmelCase , get_mid(lowerCAmelCase , lowerCAmelCase ) , get_mid(lowerCAmelCase , lowerCAmelCase ) , depth - 1 ) triangle(lowerCAmelCase , get_mid(lowerCAmelCase , lowerCAmelCase ) , get_mid(lowerCAmelCase , lowerCAmelCase ) , depth - 1 ) triangle(lowerCAmelCase , get_mid(lowerCAmelCase , lowerCAmelCase ) , get_mid(lowerCAmelCase , lowerCAmelCase ) , depth - 1 ) if __name__ == "__main__": if len(sys.argv) != 2: raise ValueError( """Correct format for using this script: """ """python fractals.py <int:depth_for_fractal>""" ) _A = turtle.Turtle() my_pen.ht() my_pen.speed(5) my_pen.pencolor("""red""") _A = [(-1_75, -1_25), (0, 1_75), (1_75, -1_25)] # vertices of triangle triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
166
1
"""simple docstring""" import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch _lowercase = '''sshleifer/bart-tiny-random''' _lowercase = '''patrickvonplaten/t5-tiny-random''' @require_torch class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[Any]: return AutoConfig.from_pretrained(A_ ) def _SCREAMING_SNAKE_CASE ( self : int ) -> List[str]: A , *A = create_student_by_copying_alternating_layers(A_ ,tempfile.mkdtemp() ,e=1 ,d=1 ) self.assertEqual(student.config.num_hidden_layers ,1 ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Dict: A , *A = create_student_by_copying_alternating_layers(A_ ,tempfile.mkdtemp() ,e=1 ,d=A_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Dict: A , *A = create_student_by_copying_alternating_layers(A_ ,tempfile.mkdtemp() ,e=1 ,d=A_ ) self.assertEqual(student.config.encoder_layers ,1 ) self.assertEqual(student.config.decoder_layers ,self.teacher_config.encoder_layers ) def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[int]: A , *A = create_student_by_copying_alternating_layers(A_ ,tempfile.mkdtemp() ,e=1 ,d=1 ) self.assertEqual(student.config.encoder_layers ,1 ) self.assertEqual(student.config.decoder_layers ,1 ) def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Tuple: with self.assertRaises(A_ ): create_student_by_copying_alternating_layers(A_ ,tempfile.mkdtemp() ,e=A_ ,d=A_ )
74
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''facebook/deit-base-distilled-patch16-224''': ( '''https://huggingface.co/facebook/deit-base-patch16-224/resolve/main/config.json''' ), # See all DeiT models at https://huggingface.co/models?filter=deit } class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: Optional[Any] = '''deit''' def __init__( self : int ,A_ : Optional[Any]=768 ,A_ : Union[str, Any]=12 ,A_ : Dict=12 ,A_ : int=3072 ,A_ : Optional[Any]="gelu" ,A_ : Dict=0.0 ,A_ : Any=0.0 ,A_ : str=0.02 ,A_ : Tuple=1e-12 ,A_ : Union[str, Any]=224 ,A_ : Optional[Any]=16 ,A_ : List[Any]=3 ,A_ : Optional[Any]=True ,A_ : Optional[int]=16 ,**A_ : Union[str, Any] ,) -> Dict: super().__init__(**A_ ) A = hidden_size A = num_hidden_layers A = num_attention_heads A = intermediate_size A = hidden_act A = hidden_dropout_prob A = attention_probs_dropout_prob A = initializer_range A = layer_norm_eps A = image_size A = patch_size A = num_channels A = qkv_bias A = encoder_stride class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: int = version.parse('''1.11''' ) @property def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> float: return 1e-4
74
1
import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __magic_name__ (__lowercase ): def __a ( self ) -> Dict: lowerCAmelCase_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_a , "embed_dim" ) ) self.parent.assertTrue(hasattr(_a , "num_heads" ) ) class __magic_name__ : def __init__( self , _a , _a=13 , _a=64 , _a=3 , _a=[16, 48, 96] , _a=[1, 3, 6] , _a=[1, 2, 10] , _a=[7, 3, 3] , _a=[4, 2, 2] , _a=[2, 1, 1] , _a=[2, 2, 2] , _a=[False, False, True] , _a=[0.0, 0.0, 0.0] , _a=0.0_2 , _a=1E-12 , _a=True , _a=True , _a=2 , ) -> List[str]: lowerCAmelCase_ = parent lowerCAmelCase_ = batch_size lowerCAmelCase_ = image_size lowerCAmelCase_ = patch_sizes lowerCAmelCase_ = patch_stride lowerCAmelCase_ = patch_padding lowerCAmelCase_ = is_training lowerCAmelCase_ = use_labels lowerCAmelCase_ = num_labels lowerCAmelCase_ = num_channels lowerCAmelCase_ = embed_dim lowerCAmelCase_ = num_heads lowerCAmelCase_ = stride_kv lowerCAmelCase_ = depth lowerCAmelCase_ = cls_token lowerCAmelCase_ = attention_drop_rate lowerCAmelCase_ = initializer_range lowerCAmelCase_ = layer_norm_eps def __a ( self ) -> Union[str, Any]: lowerCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase_ = None if self.use_labels: lowerCAmelCase_ = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase_ = self.get_config() return config, pixel_values, labels def __a ( self ) -> Dict: return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def __a ( self , _a , _a , _a ) -> Optional[Any]: lowerCAmelCase_ = CvtModel(config=_a ) model.to(_a ) model.eval() lowerCAmelCase_ = model(_a ) lowerCAmelCase_ = (self.image_size, self.image_size) lowerCAmelCase_ , lowerCAmelCase_ = image_size[0], image_size[1] for i in range(len(self.depth ) ): lowerCAmelCase_ = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) lowerCAmelCase_ = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def __a ( self , _a , _a , _a ) -> str: lowerCAmelCase_ = self.num_labels lowerCAmelCase_ = CvtForImageClassification(_a ) model.to(_a ) model.eval() lowerCAmelCase_ = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __a ( self ) -> Union[str, Any]: lowerCAmelCase_ = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = config_and_inputs lowerCAmelCase_ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __magic_name__ (__lowercase , __lowercase , unittest.TestCase ): lowerCamelCase__ = (CvtModel, CvtForImageClassification) if is_torch_available() else () lowerCamelCase__ = ( {'''feature-extraction''': CvtModel, '''image-classification''': CvtForImageClassification} if is_torch_available() else {} ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def __a ( self ) -> str: lowerCAmelCase_ = CvtModelTester(self ) lowerCAmelCase_ = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def __a ( self ) -> List[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __a ( self ) -> Tuple: return @unittest.skip(reason="Cvt does not output attentions" ) def __a ( self ) -> Tuple: pass @unittest.skip(reason="Cvt does not use inputs_embeds" ) def __a ( self ) -> Tuple: pass @unittest.skip(reason="Cvt does not support input and output embeddings" ) def __a ( self ) -> str: pass def __a ( self ) -> Dict: lowerCAmelCase_ , lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ = model_class(_a ) lowerCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase_ = [*signature.parameters.keys()] lowerCAmelCase_ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _a ) def __a ( self ) -> str: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __a ( self ) -> Union[str, Any]: def check_hidden_states_output(_a , _a , _a ): lowerCAmelCase_ = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): lowerCAmelCase_ = model(**self._prepare_for_class(_a , _a ) ) lowerCAmelCase_ = outputs.hidden_states lowerCAmelCase_ = len(self.model_tester.depth ) self.assertEqual(len(_a ) , _a ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) lowerCAmelCase_ , lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase_ = True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase_ = True check_hidden_states_output(_a , _a , _a ) def __a ( self ) -> List[Any]: lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __a ( self ) -> List[Any]: pass @slow def __a ( self ) -> int: for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ = CvtModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def A(): lowerCAmelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __magic_name__ (unittest.TestCase ): @cached_property def __a ( self ) -> int: return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __a ( self ) -> List[str]: lowerCAmelCase_ = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_a ) lowerCAmelCase_ = self.default_image_processor lowerCAmelCase_ = prepare_img() lowerCAmelCase_ = image_processor(images=_a , return_tensors="pt" ).to(_a ) # forward pass with torch.no_grad(): lowerCAmelCase_ = model(**_a ) # verify the logits lowerCAmelCase_ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _a ) lowerCAmelCase_ = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1E-4 ) )
22
def A(): return [list(range(1000 - i , -1000 - i , -1 ) ) for i in range(1000 )] lowerCamelCase__ = generate_large_matrix() lowerCamelCase__ = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def A(__a: list[list[int]] ): assert all(row == sorted(__a , reverse=__a ) for row in grid ) assert all(list(__a ) == sorted(__a , reverse=__a ) for col in zip(*__a ) ) def A(__a: list[int] ): lowerCAmelCase_ = 0 lowerCAmelCase_ = len(__a ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: lowerCAmelCase_ = (left + right) // 2 lowerCAmelCase_ = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: lowerCAmelCase_ = mid + 1 else: lowerCAmelCase_ = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__a ) def A(__a: list[list[int]] ): lowerCAmelCase_ = 0 lowerCAmelCase_ = len(grid[0] ) for i in range(len(__a ) ): lowerCAmelCase_ = find_negative_index(grid[i][:bound] ) total += bound return (len(__a ) * len(grid[0] )) - total def A(__a: list[list[int]] ): return len([number for row in grid for number in row if number < 0] ) def A(__a: list[list[int]] ): lowerCAmelCase_ = 0 for row in grid: for i, number in enumerate(__a ): if number < 0: total += len(__a ) - i break return total def A(): from timeit import timeit print("Running benchmarks" ) lowerCAmelCase_ = ( "from __main__ import count_negatives_binary_search, " "count_negatives_brute_force, count_negatives_brute_force_with_break, grid" ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): lowerCAmelCase_ = timeit(F"{func}(grid=grid)" , setup=__a , number=500 ) print(F"{func}() took {time:0.4f} seconds" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
22
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { '''camembert-base''': '''https://huggingface.co/camembert-base/resolve/main/config.json''', '''umberto-commoncrawl-cased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json''' ), '''umberto-wikipedia-uncased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json''' ), } class lowercase__ ( _UpperCAmelCase ): a_ ="""camembert""" def __init__( self , __UpperCAmelCase=30522 , __UpperCAmelCase=768 , __UpperCAmelCase=12 , __UpperCAmelCase=12 , __UpperCAmelCase=3072 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.1 , __UpperCAmelCase=512 , __UpperCAmelCase=2 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-1_2 , __UpperCAmelCase=1 , __UpperCAmelCase=0 , __UpperCAmelCase=2 , __UpperCAmelCase="absolute" , __UpperCAmelCase=True , __UpperCAmelCase=None , **__UpperCAmelCase , )-> Optional[int]: '''simple docstring''' super().__init__(pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , **__a ) lowerCAmelCase__ = vocab_size lowerCAmelCase__ = hidden_size lowerCAmelCase__ = num_hidden_layers lowerCAmelCase__ = num_attention_heads lowerCAmelCase__ = hidden_act lowerCAmelCase__ = intermediate_size lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = max_position_embeddings lowerCAmelCase__ = type_vocab_size lowerCAmelCase__ = initializer_range lowerCAmelCase__ = layer_norm_eps lowerCAmelCase__ = position_embedding_type lowerCAmelCase__ = use_cache lowerCAmelCase__ = classifier_dropout class lowercase__ ( _UpperCAmelCase ): @property def UpperCAmelCase ( self )-> int: '''simple docstring''' if self.task == "multiple-choice": lowerCAmelCase__ = {0: "batch", 1: "choice", 2: "sequence"} else: lowerCAmelCase__ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
340
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType _snake_case = logging.get_logger(__name__) class UpperCAmelCase_ ( a): lowerCamelCase__ = 'vision-encoder-decoder' lowerCamelCase__ = True def __init__( self, **__a): '''simple docstring''' super().__init__(**__a) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f"A configuraton of type {self.model_type} cannot be instantiated because " f"not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}") _lowerCAmelCase : str = kwargs.pop("encoder") _lowerCAmelCase : Any = encoder_config.pop("model_type") _lowerCAmelCase : str = kwargs.pop("decoder") _lowerCAmelCase : List[str] = decoder_config.pop("model_type") _lowerCAmelCase : Optional[Any] = AutoConfig.for_model(__a, **__a) _lowerCAmelCase : Optional[Any] = AutoConfig.for_model(__a, **__a) _lowerCAmelCase : Optional[int] = True @classmethod def snake_case__ ( cls, __a, __a, **__a): '''simple docstring''' logger.info("Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config") _lowerCAmelCase : Optional[Any] = True _lowerCAmelCase : str = True return cls(encoder=encoder_config.to_dict(), decoder=decoder_config.to_dict(), **__a) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : int = copy.deepcopy(self.__dict__) _lowerCAmelCase : List[str] = self.encoder.to_dict() _lowerCAmelCase : List[str] = self.decoder.to_dict() _lowerCAmelCase : Any = self.__class__.model_type return output class UpperCAmelCase_ ( a): lowerCamelCase__ = version.parse('1.11') @property def snake_case__ ( self): '''simple docstring''' return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ]) @property def snake_case__ ( self): '''simple docstring''' return 1E-4 @property def snake_case__ ( self): '''simple docstring''' return OrderedDict({"last_hidden_state": {0: "batch", 1: "encoder_sequence"}}) class UpperCAmelCase_ ( a): @property def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Optional[Any] = OrderedDict() _lowerCAmelCase : Any = {0: "batch", 1: "past_decoder_sequence + sequence"} _lowerCAmelCase : List[str] = {0: "batch", 1: "past_decoder_sequence + sequence"} _lowerCAmelCase : Optional[Any] = {0: "batch", 1: "encoder_sequence"} return common_inputs def snake_case__ ( self, __a, __a = -1, __a = -1, __a = False, __a = None, ): '''simple docstring''' import torch _lowerCAmelCase : Optional[Any] = OrderedDict() _lowerCAmelCase : List[str] = super().generate_dummy_inputs( __a, batch_size=__a, seq_length=__a, is_pair=__a, framework=__a) _lowerCAmelCase , _lowerCAmelCase : Optional[Any] = dummy_input["input_ids"].shape _lowerCAmelCase : str = (batch, encoder_sequence, self._config.encoder_hidden_size) _lowerCAmelCase : List[str] = dummy_input.pop("input_ids") _lowerCAmelCase : List[str] = dummy_input.pop("attention_mask") _lowerCAmelCase : Optional[int] = torch.zeros(__a) return common_inputs class UpperCAmelCase_ ( a): @property def snake_case__ ( self): '''simple docstring''' pass def snake_case__ ( self, __a): '''simple docstring''' return VisionEncoderDecoderEncoderOnnxConfig(__a) def snake_case__ ( self, __a, __a, __a = "default"): '''simple docstring''' _lowerCAmelCase : Dict = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(__a, __a)
36
0
"""simple docstring""" def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" return number | (1 << position) def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" return number & ~(1 << position) def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" return number ^ (1 << position) def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" return ((number >> position) & 1) == 1 def __UpperCAmelCase ( lowercase ,lowercase ): """simple docstring""" return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
30
"""simple docstring""" import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def __UpperCAmelCase ( lowercase=None ,lowercase=None ): """simple docstring""" return field(default_factory=lambda: default ,metadata=lowercase ) @dataclass class a : _snake_case : str = field( metadata={'help': 'The csv file to plot.'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={'help': 'Whether to plot along batch size or sequence length. Defaults to sequence length.'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={'help': 'Whether the csv file has time results or memory results. Defaults to memory results.'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={'help': 'Disable logarithmic scale when plotting'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={ 'help': 'Whether the csv file has training results or inference results. Defaults to inference results.' } , ) _snake_case : Optional[str] = field( default=lowerCAmelCase_ , metadata={'help': 'Filename under which the plot will be saved. If unused no plot is saved.'} , ) _snake_case : Optional[List[str]] = list_field( default=lowerCAmelCase_ , metadata={'help': 'List of model names that are used instead of the ones in the csv file.'} ) def __UpperCAmelCase ( lowercase ): """simple docstring""" try: int(lowercase ) return True except ValueError: return False def __UpperCAmelCase ( lowercase ): """simple docstring""" try: float(lowercase ) return True except ValueError: return False class a : def __init__( self : int , __lowerCAmelCase : Union[str, Any] ): _UpperCAmelCase = args _UpperCAmelCase = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline="""""" ) as csv_file: _UpperCAmelCase = csv.DictReader(__lowerCAmelCase ) for row in reader: _UpperCAmelCase = row["""model"""] self.result_dict[model_name]["bsz"].append(int(row["""batch_size"""] ) ) self.result_dict[model_name]["seq_len"].append(int(row["""sequence_length"""] ) ) if can_convert_to_int(row["""result"""] ): # value is not None _UpperCAmelCase = int(row["""result"""] ) elif can_convert_to_float(row["""result"""] ): # value is not None _UpperCAmelCase = float(row["""result"""] ) def lowerCAmelCase_ ( self : Optional[Any] ): _UpperCAmelCase , _UpperCAmelCase = plt.subplots() _UpperCAmelCase = """Time usage""" if self.args.is_time else """Memory usage""" _UpperCAmelCase = title_str + """ for training""" if self.args.is_train else title_str + """ for inference""" if not self.args.no_log_scale: # set logarithm scales ax.set_xscale("""log""" ) ax.set_yscale("""log""" ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): _UpperCAmelCase = sorted(set(self.result_dict[model_name]["""bsz"""] ) ) _UpperCAmelCase = sorted(set(self.result_dict[model_name]["""seq_len"""] ) ) _UpperCAmelCase = self.result_dict[model_name]["""result"""] ((_UpperCAmelCase) , (_UpperCAmelCase)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) _UpperCAmelCase = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: _UpperCAmelCase = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=__lowerCAmelCase , ) else: _UpperCAmelCase = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((_UpperCAmelCase) , (_UpperCAmelCase)) = ( ("""batch_size""", """len""") if self.args.plot_along_batch else ("""in #tokens""", """bsz""") ) _UpperCAmelCase = np.asarray(__lowerCAmelCase , __lowerCAmelCase )[: len(__lowerCAmelCase )] plt.scatter( __lowerCAmelCase , __lowerCAmelCase , label=f'''{label_model_name} - {inner_loop_label}: {inner_loop_value}''' ) plt.plot(__lowerCAmelCase , __lowerCAmelCase , """--""" ) title_str += f''' {label_model_name} vs.''' _UpperCAmelCase = title_str[:-4] _UpperCAmelCase = """Time in s""" if self.args.is_time else """Memory in MB""" # plot plt.title(__lowerCAmelCase ) plt.xlabel(__lowerCAmelCase ) plt.ylabel(__lowerCAmelCase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def __UpperCAmelCase ( ): """simple docstring""" _UpperCAmelCase = HfArgumentParser(lowercase ) _UpperCAmelCase = parser.parse_args_into_dataclasses()[0] _UpperCAmelCase = Plot(args=lowercase ) plot.plot() if __name__ == "__main__": main()
30
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=18 , __lowercase=30 , __lowercase=400 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , __lowercase=True , ) -> Dict: __UpperCamelCase :Union[str, Any] = size if size is not None else {'''shortest_edge''': 20} __UpperCamelCase :List[str] = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} __UpperCamelCase :int = parent __UpperCamelCase :Union[str, Any] = batch_size __UpperCamelCase :Tuple = num_channels __UpperCamelCase :Tuple = image_size __UpperCamelCase :List[Any] = min_resolution __UpperCamelCase :Tuple = max_resolution __UpperCamelCase :Tuple = do_resize __UpperCamelCase :Any = size __UpperCamelCase :Optional[int] = do_center_crop __UpperCamelCase :int = crop_size __UpperCamelCase :Optional[int] = do_flip_channel_order def UpperCamelCase__ ( self) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Optional[int] = MobileViTImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :int = MobileViTImageProcessingTester(self) @property def UpperCamelCase__ ( self) -> List[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Union[str, Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__lowercase , '''do_resize''')) self.assertTrue(hasattr(__lowercase , '''size''')) self.assertTrue(hasattr(__lowercase , '''do_center_crop''')) self.assertTrue(hasattr(__lowercase , '''center_crop''')) self.assertTrue(hasattr(__lowercase , '''do_flip_channel_order''')) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) __UpperCamelCase :Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCamelCase__ ( self) -> Optional[Any]: pass def UpperCamelCase__ ( self) -> List[Any]: # Initialize image_processing __UpperCamelCase :List[str] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCamelCase :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image) # Test not batched input __UpperCamelCase :int = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __UpperCamelCase :str = image_processing(__lowercase , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase__ ( self) -> Dict: # Initialize image_processing __UpperCamelCase :Union[str, Any] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCamelCase :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray) # Test not batched input __UpperCamelCase :List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __UpperCamelCase :Optional[int] = image_processing(__lowercase , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase__ ( self) -> List[str]: # Initialize image_processing __UpperCamelCase :List[Any] = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCamelCase :Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor) # Test not batched input __UpperCamelCase :Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __UpperCamelCase :Optional[int] = image_processing(__lowercase , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
43
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : int = StableUnCLIPImgaImgPipeline a__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a__ : int = frozenset([] ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = 32 __UpperCamelCase :Optional[int] = embedder_hidden_size # image encoding components __UpperCamelCase :Union[str, Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__lowercase , projection_dim=__lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) __UpperCamelCase :str = StableUnCLIPImageNormalizer(embedding_dim=__lowercase) __UpperCamelCase :Optional[int] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) __UpperCamelCase :Dict = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , )) torch.manual_seed(0) __UpperCamelCase :List[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Tuple = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0) __UpperCamelCase :List[str] = AutoencoderKL() __UpperCamelCase :Tuple = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0 , __lowercase=True) -> str: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :Union[str, Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :int = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) if pil_image: __UpperCamelCase :List[Any] = input_image * 0.5 + 0.5 __UpperCamelCase :Optional[Any] = input_image.clamp(0 , 1) __UpperCamelCase :int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __UpperCamelCase :Optional[Any] = DiffusionPipeline.numpy_to_pil(__lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Tuple = self.get_dummy_components() __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline(**__lowercase) __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) inputs.update({'''image_embeds''': None}) __UpperCamelCase :Any = sd_pipe(**__lowercase).images __UpperCamelCase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__lowercase) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Dict = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Optional[int] = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) __UpperCamelCase :Union[str, Any] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :Optional[Any] = pipe( __lowercase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) __UpperCamelCase :int = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
43
1
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase_ : Tuple = logging.get_logger(__name__) UpperCamelCase_ : int = { '''SenseTime/deformable-detr''': '''https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json''', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class _a ( __lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : str = """deformable_detr""" SCREAMING_SNAKE_CASE_ : Dict = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=None ,_SCREAMING_SNAKE_CASE=3 ,_SCREAMING_SNAKE_CASE=300 ,_SCREAMING_SNAKE_CASE=1_024 ,_SCREAMING_SNAKE_CASE=6 ,_SCREAMING_SNAKE_CASE=1_024 ,_SCREAMING_SNAKE_CASE=8 ,_SCREAMING_SNAKE_CASE=6 ,_SCREAMING_SNAKE_CASE=1_024 ,_SCREAMING_SNAKE_CASE=8 ,_SCREAMING_SNAKE_CASE=0.0 ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE="relu" ,_SCREAMING_SNAKE_CASE=256 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.0 ,_SCREAMING_SNAKE_CASE=0.0 ,_SCREAMING_SNAKE_CASE=0.0_2 ,_SCREAMING_SNAKE_CASE=1.0 ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE="sine" ,_SCREAMING_SNAKE_CASE="resnet50" ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE=4 ,_SCREAMING_SNAKE_CASE=4 ,_SCREAMING_SNAKE_CASE=4 ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE=300 ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE=1 ,_SCREAMING_SNAKE_CASE=5 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE=1 ,_SCREAMING_SNAKE_CASE=1 ,_SCREAMING_SNAKE_CASE=5 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.2_5 ,_SCREAMING_SNAKE_CASE=False ,**_SCREAMING_SNAKE_CASE ,) -> Optional[Any]: if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) _snake_case = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ): _snake_case = backbone_config.get("model_type" ) _snake_case = CONFIG_MAPPING[backbone_model_type] _snake_case = config_class.from_dict(_SCREAMING_SNAKE_CASE ) _snake_case = use_timm_backbone _snake_case = backbone_config _snake_case = num_channels _snake_case = num_queries _snake_case = max_position_embeddings _snake_case = d_model _snake_case = encoder_ffn_dim _snake_case = encoder_layers _snake_case = encoder_attention_heads _snake_case = decoder_ffn_dim _snake_case = decoder_layers _snake_case = decoder_attention_heads _snake_case = dropout _snake_case = attention_dropout _snake_case = activation_dropout _snake_case = activation_function _snake_case = init_std _snake_case = init_xavier_std _snake_case = encoder_layerdrop _snake_case = auxiliary_loss _snake_case = position_embedding_type _snake_case = backbone _snake_case = use_pretrained_backbone _snake_case = dilation # deformable attributes _snake_case = num_feature_levels _snake_case = encoder_n_points _snake_case = decoder_n_points _snake_case = two_stage _snake_case = two_stage_num_proposals _snake_case = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError("If two_stage is True, with_box_refine must be True." ) # Hungarian matcher _snake_case = class_cost _snake_case = bbox_cost _snake_case = giou_cost # Loss coefficients _snake_case = mask_loss_coefficient _snake_case = dice_loss_coefficient _snake_case = bbox_loss_coefficient _snake_case = giou_loss_coefficient _snake_case = eos_coefficient _snake_case = focal_alpha _snake_case = disable_custom_kernels super().__init__(is_encoder_decoder=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ) @property def _lowercase ( self ) -> int: return self.encoder_attention_heads @property def _lowercase ( self ) -> int: return self.d_model def _lowercase ( self ) -> Optional[int]: _snake_case = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: _snake_case = self.backbone_config.to_dict() _snake_case = self.__class__.model_type return output
142
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase_ : str = logging.get_logger(__name__) UpperCamelCase_ : Optional[Any] = { '''asapp/sew-tiny-100k''': '''https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json''', # See all SEW models at https://huggingface.co/models?filter=sew } class _a ( __lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : Dict = """sew""" def __init__( self ,_SCREAMING_SNAKE_CASE=32 ,_SCREAMING_SNAKE_CASE=768 ,_SCREAMING_SNAKE_CASE=12 ,_SCREAMING_SNAKE_CASE=12 ,_SCREAMING_SNAKE_CASE=3_072 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE="gelu" ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.0 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.1 ,_SCREAMING_SNAKE_CASE=0.0_2 ,_SCREAMING_SNAKE_CASE=1e-5 ,_SCREAMING_SNAKE_CASE="group" ,_SCREAMING_SNAKE_CASE="gelu" ,_SCREAMING_SNAKE_CASE=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) ,_SCREAMING_SNAKE_CASE=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) ,_SCREAMING_SNAKE_CASE=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE=128 ,_SCREAMING_SNAKE_CASE=16 ,_SCREAMING_SNAKE_CASE=True ,_SCREAMING_SNAKE_CASE=0.0_5 ,_SCREAMING_SNAKE_CASE=10 ,_SCREAMING_SNAKE_CASE=2 ,_SCREAMING_SNAKE_CASE=0.0 ,_SCREAMING_SNAKE_CASE=10 ,_SCREAMING_SNAKE_CASE=0 ,_SCREAMING_SNAKE_CASE="mean" ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE=False ,_SCREAMING_SNAKE_CASE=256 ,_SCREAMING_SNAKE_CASE=0 ,_SCREAMING_SNAKE_CASE=1 ,_SCREAMING_SNAKE_CASE=2 ,**_SCREAMING_SNAKE_CASE ,) -> str: super().__init__(**_SCREAMING_SNAKE_CASE ,pad_token_id=_SCREAMING_SNAKE_CASE ,bos_token_id=_SCREAMING_SNAKE_CASE ,eos_token_id=_SCREAMING_SNAKE_CASE ) _snake_case = hidden_size _snake_case = feat_extract_norm _snake_case = feat_extract_activation _snake_case = list(_SCREAMING_SNAKE_CASE ) _snake_case = list(_SCREAMING_SNAKE_CASE ) _snake_case = list(_SCREAMING_SNAKE_CASE ) _snake_case = conv_bias _snake_case = num_conv_pos_embeddings _snake_case = num_conv_pos_embedding_groups _snake_case = len(self.conv_dim ) _snake_case = num_hidden_layers _snake_case = intermediate_size _snake_case = squeeze_factor _snake_case = hidden_act _snake_case = num_attention_heads _snake_case = hidden_dropout _snake_case = attention_dropout _snake_case = activation_dropout _snake_case = feat_proj_dropout _snake_case = final_dropout _snake_case = layerdrop _snake_case = layer_norm_eps _snake_case = initializer_range _snake_case = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect." "It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`," f"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" f"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _snake_case = apply_spec_augment _snake_case = mask_time_prob _snake_case = mask_time_length _snake_case = mask_time_min_masks _snake_case = mask_feature_prob _snake_case = mask_feature_length _snake_case = mask_feature_min_masks # ctc loss _snake_case = ctc_loss_reduction _snake_case = ctc_zero_infinity # sequence classification _snake_case = use_weighted_layer_sum _snake_case = classifier_proj_size @property def _lowercase ( self ) -> Optional[Any]: return functools.reduce(operator.mul ,self.conv_stride ,1 )
142
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor UpperCAmelCase__ : int = logging.get_logger(__name__) class lowerCAmelCase_ (_lowerCAmelCase ): """simple docstring""" def __init__(self , *SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: """simple docstring""" warnings.warn( """The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use MobileViTImageProcessor instead.""" , _lowerCamelCase , ) super().__init__(*_lowerCamelCase , **_lowerCamelCase )
25
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class snake_case ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : Any, _lowerCamelCase : Optional[int] ): '''simple docstring''' for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''], model_result['''ss'''] ): __A = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(_lowerCamelCase ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' __A = '''sgugger/tiny-distilbert-classification''' __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, only_pretrain_model=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Any ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, torchscript=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == '''cpu''', '''Cant do half precision''' ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, fpaa=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Dict ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = AutoConfig.from_pretrained(_lowerCamelCase ) # set architectures equal to `None` __A = None __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase, configs=[config] ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : int ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == '''cpu''', '''Can\'t do half precision''' ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], fpaa=_lowerCamelCase, multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _SCREAMING_SNAKE_CASE ( self : str ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = AutoConfig.from_pretrained(_lowerCamelCase ) __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase, configs=[config] ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ): '''simple docstring''' __A = '''sshleifer/tinier_bart''' __A = AutoConfig.from_pretrained(_lowerCamelCase ) __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase, configs=[config] ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' __A = AutoConfig.from_pretrained(_lowerCamelCase ) __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase, configs=[config] ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _SCREAMING_SNAKE_CASE ( self : Dict ): '''simple docstring''' __A = '''sshleifer/tinier_bart''' __A = AutoConfig.from_pretrained(_lowerCamelCase ) __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase, configs=[config] ) __A = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, save_to_csv=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], inference_time_csv_file=os.path.join(_lowerCamelCase, '''inf_time.csv''' ), train_memory_csv_file=os.path.join(_lowerCamelCase, '''train_mem.csv''' ), inference_memory_csv_file=os.path.join(_lowerCamelCase, '''inf_mem.csv''' ), train_time_csv_file=os.path.join(_lowerCamelCase, '''train_time.csv''' ), env_info_csv_file=os.path.join(_lowerCamelCase, '''env.csv''' ), multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) benchmark.run() self.assertTrue(Path(os.path.join(_lowerCamelCase, '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase, '''train_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase, '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase, '''train_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowerCamelCase, '''env.csv''' ) ).exists() ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): '''simple docstring''' __A = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(_lowerCamelCase : List[Any] ): self.assertTrue(hasattr(_lowerCamelCase, '''sequential''' ) ) self.assertTrue(hasattr(_lowerCamelCase, '''cumulative''' ) ) self.assertTrue(hasattr(_lowerCamelCase, '''current''' ) ) self.assertTrue(hasattr(_lowerCamelCase, '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: __A = PyTorchBenchmarkArguments( models=[MODEL_ID], training=_lowerCamelCase, inference=_lowerCamelCase, sequence_lengths=[8], batch_sizes=[1], log_filename=os.path.join(_lowerCamelCase, '''log.txt''' ), log_print=_lowerCamelCase, trace_memory_line_by_line=_lowerCamelCase, multi_process=_lowerCamelCase, ) __A = PyTorchBenchmark(_lowerCamelCase ) __A = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(_lowerCamelCase, '''log.txt''' ) ).exists() )
266
0
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch _lowerCamelCase : List[str] = True except ImportError: _lowerCamelCase : Dict = False try: from torch.hub import _get_torch_home _lowerCamelCase : List[str] = _get_torch_home() except ImportError: _lowerCamelCase : List[Any] = os.path.expanduser( os.getenv("""TORCH_HOME""", os.path.join(os.getenv("""XDG_CACHE_HOME""", """~/.cache"""), """torch""")) ) _lowerCamelCase : str = os.path.join(torch_cache_home, """transformers""") _lowerCamelCase : Any = """https://cdn.huggingface.co""" _lowerCamelCase : Optional[int] = """https://s3.amazonaws.com/models.huggingface.co/bert""" _lowerCamelCase : Dict = """/""".join(str(Path(__file__).resolve()).split("""/""")[:-1]) _lowerCamelCase : Tuple = os.path.join(PATH, """config.yaml""") _lowerCamelCase : Dict = os.path.join(PATH, """attributes.txt""") _lowerCamelCase : Union[str, Any] = os.path.join(PATH, """objects.txt""") _lowerCamelCase : Any = os.getenv("""PYTORCH_PRETRAINED_BERT_CACHE""", default_cache_path) _lowerCamelCase : Union[str, Any] = os.getenv("""PYTORCH_TRANSFORMERS_CACHE""", PYTORCH_PRETRAINED_BERT_CACHE) _lowerCamelCase : int = os.getenv("""TRANSFORMERS_CACHE""", PYTORCH_TRANSFORMERS_CACHE) _lowerCamelCase : Optional[Any] = """pytorch_model.bin""" _lowerCamelCase : Union[str, Any] = """config.yaml""" def SCREAMING_SNAKE_CASE ( lowercase_=OBJECTS , lowercase_=ATTRIBUTES ) -> Tuple: """simple docstring""" A__ = [] with open(lowercase_ ) as f: for object in f.readlines(): vg_classes.append(object.split(''',''' )[0].lower().strip() ) A__ = [] with open(lowercase_ ) as f: for object in f.readlines(): vg_attrs.append(object.split(''',''' )[0].lower().strip() ) return vg_classes, vg_attrs def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Optional[Any]: """simple docstring""" A__ = OrderedDict() with open(lowercase_ , '''rb''' ) as f: A__ = pkl.load(lowercase_ )['''model'''] for k in copy.deepcopy(list(ckp.keys() ) ): A__ = ckp.pop(lowercase_ ) if isinstance(lowercase_ , np.ndarray ): A__ = torch.tensor(lowercase_ ) else: assert isinstance(lowercase_ , torch.tensor ), type(lowercase_ ) A__ = v return r class UpperCamelCase_ : '''simple docstring''' UpperCAmelCase__ = {} def __init__( self : Tuple , UpperCAmelCase__ : dict , UpperCAmelCase__ : str = "root" , UpperCAmelCase__ : List[Any]=0) ->Union[str, Any]: '''simple docstring''' A__ = name A__ = level A__ = {} for k, v in dictionary.items(): if v is None: raise ValueError() A__ = copy.deepcopy(UpperCAmelCase__) A__ = copy.deepcopy(UpperCAmelCase__) if isinstance(UpperCAmelCase__ , UpperCAmelCase__): A__ = Config(UpperCAmelCase__ , name=UpperCAmelCase__ , level=level + 1) A__ = v setattr(self , UpperCAmelCase__ , UpperCAmelCase__) A__ = d def __repr__( self : Tuple) ->Optional[int]: '''simple docstring''' return str(list((self._pointer.keys()))) def __setattr__( self : Tuple , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple) ->int: '''simple docstring''' A__ = val A__ = val A__ = key.split('''.''') A__ = len(UpperCAmelCase__) - 1 A__ = self._pointer if len(UpperCAmelCase__) > 1: for i, l in enumerate(UpperCAmelCase__): if hasattr(self , UpperCAmelCase__) and isinstance(getattr(self , UpperCAmelCase__) , UpperCAmelCase__): setattr(getattr(self , UpperCAmelCase__) , '''.'''.join(levels[i:]) , UpperCAmelCase__) if l == last_level: A__ = val else: A__ = pointer[l] def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->List[Any]: '''simple docstring''' return self._pointer def SCREAMING_SNAKE_CASE ( self : Tuple , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : List[str]) ->str: '''simple docstring''' with open(f"""{file_name}""" , '''w''') as stream: dump(UpperCAmelCase__ , UpperCAmelCase__) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[Any]) ->List[Any]: '''simple docstring''' with open(f"""{file_name}""" , '''w''') as stream: json.dump(UpperCAmelCase__ , UpperCAmelCase__) @staticmethod def SCREAMING_SNAKE_CASE ( UpperCAmelCase__ : Tuple) ->Union[str, Any]: '''simple docstring''' with open(UpperCAmelCase__) as stream: A__ = load(UpperCAmelCase__ , Loader=UpperCAmelCase__) return data def __str__( self : Union[str, Any]) ->List[Any]: '''simple docstring''' A__ = ''' ''' if self._name != "root": A__ = f"""{t * (self._level-1)}{self._name}:\n""" else: A__ = '''''' A__ = self._level for i, (k, v) in enumerate(self._pointer.items()): if isinstance(UpperCAmelCase__ , UpperCAmelCase__): r += f"""{t * (self._level)}{v}\n""" self._level += 1 else: r += f"""{t * (self._level)}{k}: {v} ({type(UpperCAmelCase__).__name__})\n""" A__ = level return r[:-1] @classmethod def SCREAMING_SNAKE_CASE ( cls : int , UpperCAmelCase__ : str , **UpperCAmelCase__ : List[Any]) ->Dict: '''simple docstring''' A__ , A__ = cls.get_config_dict(UpperCAmelCase__ , **UpperCAmelCase__) return cls(UpperCAmelCase__) @classmethod def SCREAMING_SNAKE_CASE ( cls : List[str] , UpperCAmelCase__ : str , **UpperCAmelCase__ : Tuple) ->Optional[Any]: '''simple docstring''' A__ = kwargs.pop('''cache_dir''' , UpperCAmelCase__) A__ = kwargs.pop('''force_download''' , UpperCAmelCase__) A__ = kwargs.pop('''resume_download''' , UpperCAmelCase__) A__ = kwargs.pop('''proxies''' , UpperCAmelCase__) A__ = kwargs.pop('''local_files_only''' , UpperCAmelCase__) if os.path.isdir(UpperCAmelCase__): A__ = os.path.join(UpperCAmelCase__ , UpperCAmelCase__) elif os.path.isfile(UpperCAmelCase__) or is_remote_url(UpperCAmelCase__): A__ = pretrained_model_name_or_path else: A__ = hf_bucket_url(UpperCAmelCase__ , filename=UpperCAmelCase__ , use_cdn=UpperCAmelCase__) try: # Load from URL or cache if already cached A__ = cached_path( UpperCAmelCase__ , cache_dir=UpperCAmelCase__ , force_download=UpperCAmelCase__ , proxies=UpperCAmelCase__ , resume_download=UpperCAmelCase__ , local_files_only=UpperCAmelCase__ , ) # Load config dict if resolved_config_file is None: raise EnvironmentError A__ = Config.load_yaml(UpperCAmelCase__) except EnvironmentError: A__ = '''Can\'t load config for''' raise EnvironmentError(UpperCAmelCase__) if resolved_config_file == config_file: print('''loading configuration file from path''') else: print('''loading configuration file cache''') return Config.load_yaml(UpperCAmelCase__), kwargs def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Tuple: """simple docstring""" A__ = torch.load('''dump.pt''' , map_location=in_tensor.device ) A__ = in_tensor.numpy() A__ = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(lowercase_ , lowercase_ , rtol=0.01 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(lowercase_ , lowercase_ , rtol=0.01 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %""" " element-wise mismatch" ) raise Exception('''tensors are all good''' ) # Hugging face functions below def SCREAMING_SNAKE_CASE ( lowercase_ ) -> List[str]: """simple docstring""" A__ = urlparse(lowercase_ ) return parsed.scheme in ("http", "https") def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_=True ) -> str: """simple docstring""" A__ = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX A__ = '''/''' not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_=None , lowercase_=0 , lowercase_=None , ) -> List[str]: """simple docstring""" A__ = '''python/{}'''.format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(lowercase_ , lowercase_ ): ua += "; " + "; ".join('''{}/{}'''.format(lowercase_ , lowercase_ ) for k, v in user_agent.items() ) elif isinstance(lowercase_ , lowercase_ ): ua += "; " + user_agent A__ = {'''user-agent''': ua} if resume_size > 0: A__ = '''bytes=%d-''' % (resume_size,) A__ = requests.get(lowercase_ , stream=lowercase_ , proxies=lowercase_ , headers=lowercase_ ) if response.status_code == 416: # Range not satisfiable return A__ = response.headers.get('''Content-Length''' ) A__ = resume_size + int(lowercase_ ) if content_length is not None else None A__ = tqdm( unit='''B''' , unit_scale=lowercase_ , total=lowercase_ , initial=lowercase_ , desc='''Downloading''' , ) for chunk in response.iter_content(chunk_size=1_024 ): if chunk: # filter out keep-alive new chunks progress.update(len(lowercase_ ) ) temp_file.write(lowercase_ ) progress.close() def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_=None , lowercase_=False , lowercase_=None , lowercase_=10 , lowercase_=False , lowercase_=None , lowercase_=False , ) -> List[str]: """simple docstring""" if cache_dir is None: A__ = TRANSFORMERS_CACHE if isinstance(lowercase_ , lowercase_ ): A__ = str(lowercase_ ) os.makedirs(lowercase_ , exist_ok=lowercase_ ) A__ = None if not local_files_only: try: A__ = requests.head(lowercase_ , allow_redirects=lowercase_ , proxies=lowercase_ , timeout=lowercase_ ) if response.status_code == 200: A__ = response.headers.get('''ETag''' ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass A__ = url_to_filename(lowercase_ , lowercase_ ) # get cache path to put the file A__ = os.path.join(lowercase_ , lowercase_ ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(lowercase_ ): return cache_path else: A__ = [ file for file in fnmatch.filter(os.listdir(lowercase_ ) , filename + '''.*''' ) if not file.endswith('''.json''' ) and not file.endswith('''.lock''' ) ] if len(lowercase_ ) > 0: return os.path.join(lowercase_ , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( '''Cannot find the requested files in the cached path and outgoing traffic has been''' ''' disabled. To enable model look-ups and downloads online, set \'local_files_only\'''' ''' to False.''' ) return None # From now on, etag is not None. if os.path.exists(lowercase_ ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. A__ = cache_path + '''.lock''' with FileLock(lowercase_ ): # If the download just completed while the lock was activated. if os.path.exists(lowercase_ ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: A__ = cache_path + '''.incomplete''' @contextmanager def _resumable_file_manager(): with open(lowercase_ , '''a+b''' ) as f: yield f A__ = _resumable_file_manager if os.path.exists(lowercase_ ): A__ = os.stat(lowercase_ ).st_size else: A__ = 0 else: A__ = partial(tempfile.NamedTemporaryFile , dir=lowercase_ , delete=lowercase_ ) A__ = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( '''%s not found in cache or force_download set to True, downloading to %s''' , lowercase_ , temp_file.name , ) http_get( lowercase_ , lowercase_ , proxies=lowercase_ , resume_size=lowercase_ , user_agent=lowercase_ , ) os.replace(temp_file.name , lowercase_ ) A__ = {'''url''': url, '''etag''': etag} A__ = cache_path + '''.json''' with open(lowercase_ , '''w''' ) as meta_file: json.dump(lowercase_ , lowercase_ ) return cache_path def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_=None ) -> Any: """simple docstring""" A__ = url.encode('''utf-8''' ) A__ = shaaaa(lowercase_ ) A__ = url_hash.hexdigest() if etag: A__ = etag.encode('''utf-8''' ) A__ = shaaaa(lowercase_ ) filename += "." + etag_hash.hexdigest() if url.endswith('''.h5''' ): filename += ".h5" return filename def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_=None , lowercase_=False , lowercase_=None , lowercase_=False , lowercase_=None , lowercase_=False , lowercase_=False , lowercase_=False , ) -> Tuple: """simple docstring""" if cache_dir is None: A__ = TRANSFORMERS_CACHE if isinstance(lowercase_ , lowercase_ ): A__ = str(lowercase_ ) if isinstance(lowercase_ , lowercase_ ): A__ = str(lowercase_ ) if is_remote_url(lowercase_ ): # URL, so get it from the cache (downloading if necessary) A__ = get_from_cache( lowercase_ , cache_dir=lowercase_ , force_download=lowercase_ , proxies=lowercase_ , resume_download=lowercase_ , user_agent=lowercase_ , local_files_only=lowercase_ , ) elif os.path.exists(lowercase_ ): # File, and it exists. A__ = url_or_filename elif urlparse(lowercase_ ).scheme == "": # File, but it doesn't exist. raise EnvironmentError('''file {} not found'''.format(lowercase_ ) ) else: # Something unknown raise ValueError('''unable to parse {} as a URL or as a local path'''.format(lowercase_ ) ) if extract_compressed_file: if not is_zipfile(lowercase_ ) and not tarfile.is_tarfile(lowercase_ ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" A__ , A__ = os.path.split(lowercase_ ) A__ = output_file.replace('''.''' , '''-''' ) + '''-extracted''' A__ = os.path.join(lowercase_ , lowercase_ ) if os.path.isdir(lowercase_ ) and os.listdir(lowercase_ ) and not force_extract: return output_path_extracted # Prevent parallel extractions A__ = output_path + '''.lock''' with FileLock(lowercase_ ): shutil.rmtree(lowercase_ , ignore_errors=lowercase_ ) os.makedirs(lowercase_ ) if is_zipfile(lowercase_ ): with ZipFile(lowercase_ , '''r''' ) as zip_file: zip_file.extractall(lowercase_ ) zip_file.close() elif tarfile.is_tarfile(lowercase_ ): A__ = tarfile.open(lowercase_ ) tar_file.extractall(lowercase_ ) tar_file.close() else: raise EnvironmentError('''Archive format of {} could not be identified'''.format(lowercase_ ) ) return output_path_extracted return output_path def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_="," ) -> Union[str, Any]: """simple docstring""" assert isinstance(lowercase_ , lowercase_ ) if os.path.isfile(lowercase_ ): with open(lowercase_ ) as f: A__ = eval(f.read() ) else: A__ = requests.get(lowercase_ ) try: A__ = requests.json() except Exception: A__ = req.content.decode() assert data is not None, "could not connect" try: A__ = eval(lowercase_ ) except Exception: A__ = data.split('''\n''' ) req.close() return data def SCREAMING_SNAKE_CASE ( lowercase_ ) -> int: """simple docstring""" A__ = requests.get(lowercase_ ) A__ = np.array(Image.open(BytesIO(response.content ) ) ) return img def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Any: """simple docstring""" A__ = url.split('''/''' )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(lowercase_ ) with open(lowercase_ , '''rb''' ) as stream: A__ = pkl.load(lowercase_ ) A__ = weights.pop('''model''' ) A__ = {} for k, v in model.items(): A__ = torch.from_numpy(lowercase_ ) if "running_var" in k: A__ = torch.tensor([0] ) A__ = k.replace('''running_var''' , '''num_batches_tracked''' ) A__ = zero return new def SCREAMING_SNAKE_CASE ( ) -> List[str]: """simple docstring""" print(f"""{os.path.abspath(os.path.join(lowercase_ , os.pardir ) )}/demo.ipynb""" ) def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_="RGB" ) -> Optional[Any]: """simple docstring""" assert isinstance(lowercase_ , lowercase_ ) if os.path.isfile(lowercase_ ): A__ = cva.imread(lowercase_ ) else: A__ = get_image_from_url(lowercase_ ) assert img is not None, f"""could not connect to: {im}""" A__ = cva.cvtColor(lowercase_ , cva.COLOR_BGR2RGB ) if input_format == "RGB": A__ = img[:, :, ::-1] return img def SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_=1 ) -> Tuple: """simple docstring""" return (images[i : i + batch] for i in range(0 , len(lowercase_ ) , lowercase_ ))
231
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class UpperCamelCase_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE ( self : Dict) ->Tuple: '''simple docstring''' A__ = [[1, 2, 4], [1, 2, 3, 4]] A__ = DisjunctiveConstraint(UpperCAmelCase__) self.assertTrue(isinstance(dc.token_ids , UpperCAmelCase__)) with self.assertRaises(UpperCAmelCase__): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]])) with self.assertRaises(UpperCAmelCase__): DisjunctiveConstraint([torch.LongTensor([1, 2, 4]), torch.LongTensor([1, 2, 3, 4, 5])]) def SCREAMING_SNAKE_CASE ( self : int) ->str: '''simple docstring''' A__ = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(UpperCAmelCase__): DisjunctiveConstraint(UpperCAmelCase__) # fails here def SCREAMING_SNAKE_CASE ( self : Any) ->str: '''simple docstring''' A__ = [[1, 2, 3], [1, 2, 4]] A__ = DisjunctiveConstraint(UpperCAmelCase__) A__ , A__ , A__ = dc.update(1) A__ = stepped is True and completed is False and reset is False self.assertTrue(UpperCAmelCase__) self.assertTrue(not dc.completed) self.assertTrue(dc.current_seq == [1]) A__ , A__ , A__ = dc.update(2) A__ = stepped is True and completed is False and reset is False self.assertTrue(UpperCAmelCase__) self.assertTrue(not dc.completed) self.assertTrue(dc.current_seq == [1, 2]) A__ , A__ , A__ = dc.update(3) A__ = stepped is True and completed is True and reset is False self.assertTrue(UpperCAmelCase__) self.assertTrue(dc.completed) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3]) def SCREAMING_SNAKE_CASE ( self : Optional[Any]) ->Tuple: '''simple docstring''' A__ = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] A__ = DisjunctiveConstraint(UpperCAmelCase__) A__ , A__ , A__ = dc.update(1) self.assertTrue(not dc.completed) self.assertTrue(dc.current_seq == [1]) A__ , A__ , A__ = dc.update(2) self.assertTrue(not dc.completed) self.assertTrue(dc.current_seq == [1, 2]) A__ , A__ , A__ = dc.update(4) self.assertTrue(not dc.completed) self.assertTrue(dc.current_seq == [1, 2, 4]) A__ , A__ , A__ = dc.update(5) self.assertTrue(dc.completed) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5]) dc.reset() A__ , A__ , A__ = dc.update(1) self.assertTrue(not dc.completed) self.assertTrue(dc.remaining() == 3) self.assertTrue(dc.current_seq == [1]) A__ , A__ , A__ = dc.update(2) self.assertTrue(not dc.completed) self.assertTrue(dc.remaining() == 2) self.assertTrue(dc.current_seq == [1, 2]) A__ , A__ , A__ = dc.update(5) self.assertTrue(dc.completed) # Completed! self.assertTrue(dc.remaining() == 0) self.assertTrue(dc.current_seq == [1, 2, 5])
231
1
import math __lowerCAmelCase : int = 10 __lowerCAmelCase : str = 7 __lowerCAmelCase : Union[str, Any] = BALLS_PER_COLOUR * NUM_COLOURS def UpperCAmelCase_ ( __lowerCAmelCase = 20 ) -> Optional[int]: __lowercase : List[Any] = math.comb(__lowerCAmelCase , __lowerCAmelCase ) __lowercase : Any = math.comb(NUM_BALLS - BALLS_PER_COLOUR , __lowerCAmelCase ) __lowercase : List[str] = NUM_COLOURS * (1 - missing_colour / total) return F'{result:.9f}' if __name__ == "__main__": print(solution(20))
156
'''simple docstring''' from __future__ import annotations from math import ceil, floor, sqrt def lowerCamelCase ( lowerCAmelCase : int = 200_0000 ): """simple docstring""" __magic_name__ : list[int] = [0] __magic_name__ : int for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target __magic_name__ : int = 0 # the area corresponding to the grid that gives the product closest to target __magic_name__ : int = 0 # an estimate of b, using the quadratic formula __magic_name__ : float # the largest integer less than b_estimate __magic_name__ : int # the largest integer less than b_estimate __magic_name__ : int # the triangle number corresponding to b_floor __magic_name__ : int # the triangle number corresponding to b_ceil __magic_name__ : int for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): __magic_name__ : Dict = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 __magic_name__ : List[Any] = floor(lowerCAmelCase ) __magic_name__ : Dict = ceil(lowerCAmelCase ) __magic_name__ : Any = triangle_numbers[b_floor] __magic_name__ : Optional[int] = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): __magic_name__ : Any = triangle_b_first_guess * triangle_a __magic_name__ : Any = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): __magic_name__ : List[str] = triangle_b_second_guess * triangle_a __magic_name__ : Optional[int] = idx_a * b_ceil return area if __name__ == "__main__": print(F'{solution() = }')
331
0
'''simple docstring''' import logging from transformers import PretrainedConfig A__ : Optional[int] =logging.getLogger(__name__) A__ : Union[str, Any] ={ '''bertabs-finetuned-cnndm''': '''https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json''', } class UpperCAmelCase ( snake_case_ ): _lowercase: List[Any] = '''bertabs''' def __init__( self : int , __snake_case : int=3_05_22 , __snake_case : Any=5_12 , __snake_case : int=6 , __snake_case : List[Any]=5_12 , __snake_case : Optional[int]=8 , __snake_case : Union[str, Any]=5_12 , __snake_case : List[str]=0.2 , __snake_case : Optional[int]=6 , __snake_case : Dict=7_68 , __snake_case : str=8 , __snake_case : List[Any]=20_48 , __snake_case : int=0.2 , **__snake_case : str , ) -> Dict: super().__init__(**__snake_case ) _lowerCAmelCase = vocab_size _lowerCAmelCase = max_pos _lowerCAmelCase = enc_layers _lowerCAmelCase = enc_hidden_size _lowerCAmelCase = enc_heads _lowerCAmelCase = enc_ff_size _lowerCAmelCase = enc_dropout _lowerCAmelCase = dec_layers _lowerCAmelCase = dec_hidden_size _lowerCAmelCase = dec_heads _lowerCAmelCase = dec_ff_size _lowerCAmelCase = dec_dropout
220
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets A__ : List[str] =datasets.logging.get_logger(__name__) A__ : List[Any] ='''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' A__ : List[str] ='''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' A__ : List[Any] =''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase=False , lowerCAmelCase=False , lowerCAmelCase=True , lowerCAmelCase=False , lowerCAmelCase="dummy_doc" ): """simple docstring""" _lowerCAmelCase = {doc: key_lines} _lowerCAmelCase = {doc: sys_lines} _lowerCAmelCase = {} _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase = 0 _lowerCAmelCase , _lowerCAmelCase = reader.get_doc_mentions(lowerCAmelCase , key_doc_lines[doc] , lowerCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase = reader.set_annotated_parse_trees(lowerCAmelCase , key_doc_lines[doc] , lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase , _lowerCAmelCase = reader.get_doc_mentions(lowerCAmelCase , sys_doc_lines[doc] , lowerCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: _lowerCAmelCase = reader.set_annotated_parse_trees(lowerCAmelCase , key_doc_lines[doc] , lowerCAmelCase , lowerCAmelCase ) if remove_nested: _lowerCAmelCase , _lowerCAmelCase = reader.remove_nested_coref_mentions(lowerCAmelCase , lowerCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _lowerCAmelCase , _lowerCAmelCase = reader.remove_nested_coref_mentions(lowerCAmelCase , lowerCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _lowerCAmelCase = reader.get_mention_assignments(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = reader.get_mention_assignments(lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( """Number of removed nested coreferring mentions in the key """ f"annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}" ) logger.info( """Number of resulting singleton clusters in the key """ f"annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}" ) if not keep_singletons: logger.info( f"{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system " """files, respectively""" ) return doc_coref_infos def UpperCamelCase__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = get_coref_infos(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) _lowerCAmelCase = {} _lowerCAmelCase = 0 _lowerCAmelCase = 0 for name, metric in metrics: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = evaluator.evaluate_documents(lowerCAmelCase , lowerCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"{name}/recall": recall, f"{name}/precision": precision, f"{name}/f1": fa} ) logger.info( name.ljust(10 ) , f"Recall: {recall * 1_00:.2f}" , f" Precision: {precision * 1_00:.2f}" , f" F1: {fa * 1_00:.2f}" , ) if conll_subparts_num == 3: _lowerCAmelCase = (conll / 3) * 1_00 logger.info(f"CoNLL score: {conll:.2f}" ) output_scores.update({"""conll_score""": conll} ) return output_scores def UpperCamelCase__ ( lowerCAmelCase ): """simple docstring""" _lowerCAmelCase = False for line in key_lines: if not line.startswith("""#""" ): if len(line.split() ) > 6: _lowerCAmelCase = line.split()[5] if not parse_col == "-": _lowerCAmelCase = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): def lowercase__ ( self : str ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Sequence(datasets.Value("""string""" ) ), """references""": datasets.Sequence(datasets.Value("""string""" ) ), } ) , codebase_urls=["""https://github.com/ns-moosavi/coval"""] , reference_urls=[ """https://github.com/ns-moosavi/coval""", """https://www.aclweb.org/anthology/P16-1060""", """http://www.conll.cemantix.org/2012/data.html""", ] , ) def lowercase__ ( self : List[Any] , __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : Dict=True , __snake_case : List[str]=False , __snake_case : List[Any]=False , __snake_case : Dict=False ) -> Union[str, Any]: _lowerCAmelCase = [ ("""mentions""", evaluator.mentions), ("""muc""", evaluator.muc), ("""bcub""", evaluator.b_cubed), ("""ceafe""", evaluator.ceafe), ("""lea""", evaluator.lea), ] if min_span: _lowerCAmelCase = util.check_gold_parse_annotation(__snake_case ) if not has_gold_parse: raise NotImplementedError("""References should have gold parse annotation to use 'min_span'.""" ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _lowerCAmelCase = evaluate( key_lines=__snake_case , sys_lines=__snake_case , metrics=__snake_case , NP_only=__snake_case , remove_nested=__snake_case , keep_singletons=__snake_case , min_span=__snake_case , ) return score
220
1
import torch from diffusers import StableDiffusionPipeline _SCREAMING_SNAKE_CASE = "path-to-your-trained-model" _SCREAMING_SNAKE_CASE = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('cuda') _SCREAMING_SNAKE_CASE = "A photo of sks dog in a bucket" _SCREAMING_SNAKE_CASE = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save('dog-bucket.png')
180
def UpperCAmelCase__ (UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_=False ): """simple docstring""" if isinstance(UpperCamelCase_ ,UpperCamelCase_ ) and isinstance(UpperCamelCase_ ,UpperCamelCase_ ): snake_case = len(set_a.intersection(UpperCamelCase_ ) ) if alternative_union: snake_case = len(UpperCamelCase_ ) + len(UpperCamelCase_ ) else: snake_case = len(set_a.union(UpperCamelCase_ ) ) return intersection / union if isinstance(UpperCamelCase_ ,(list, tuple) ) and isinstance(UpperCamelCase_ ,(list, tuple) ): snake_case = [element for element in set_a if element in set_b] if alternative_union: snake_case = len(UpperCamelCase_ ) + len(UpperCamelCase_ ) return len(UpperCamelCase_ ) / union else: snake_case = set_a + [element for element in set_b if element not in set_a] return len(UpperCamelCase_ ) / len(UpperCamelCase_ ) return len(UpperCamelCase_ ) / len(UpperCamelCase_ ) return None if __name__ == "__main__": _SCREAMING_SNAKE_CASE : Optional[int] = {"a", "b", "c", "d", "e"} _SCREAMING_SNAKE_CASE : List[str] = {"c", "d", "e", "f", "h", "i"} print(jaccard_similarity(set_a, set_b))
127
0
def SCREAMING_SNAKE_CASE_ ( __A : str , __A : list[str] ) -> str: """simple docstring""" a_ : Union[str, Any] = '' for word_or_phrase in separated: if not isinstance(__A , __A ): raise Exception('join() accepts only strings to be joined' ) joined += word_or_phrase + separator return joined.strip(__A ) if __name__ == "__main__": from doctest import testmod testmod()
120
def SCREAMING_SNAKE_CASE_ ( __A : int ) -> int: """simple docstring""" if not isinstance(__A , __A ): raise ValueError('Input must be an integer' ) if input_num <= 0: raise ValueError('Input must be positive' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divisor == 0 ) if __name__ == "__main__": import doctest doctest.testmod()
120
1
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class a__ : A = 42 A = None # Automatically constructed A = "dict" A = None A = field(default='Translation' , init=A__ , repr=A__ ) def __call__( self : Tuple ): """simple docstring""" return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def __UpperCamelCase ( self : Any ): """simple docstring""" from .features import Value return {k: Value("string" ) for k in sorted(self.languages )} @dataclass class a__ : A = None A = None A = None # Automatically constructed A = "dict" A = None A = field(default='TranslationVariableLanguages' , init=A__ , repr=A__ ) def __UpperCamelCase ( self : List[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = sorted(set(self.languages ) ) if self.languages else None SCREAMING_SNAKE_CASE_ : Optional[Any] = len(self.languages ) if self.languages else None def __call__( self : Dict ): """simple docstring""" return pa.struct({"language": pa.list_(pa.string() ), "translation": pa.list_(pa.string() )} ) def __UpperCamelCase ( self : List[Any],_A : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = set(self.languages ) if self.languages and set(_A ) - lang_set: raise ValueError( F'Some languages in example ({", ".join(sorted(set(_A ) - lang_set ) )}) are not in valid set ({", ".join(_A )}).' ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. SCREAMING_SNAKE_CASE_ : Union[str, Any] = [] for lang, text in translation_dict.items(): if isinstance(_A,_A ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Optional[int] = zip(*sorted(_A ) ) return {"language": languages, "translation": translations} def __UpperCamelCase ( self : Dict ): """simple docstring""" from .features import Sequence, Value return { "language": Sequence(Value("string" ) ), "translation": Sequence(Value("string" ) ), }
18
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowerCamelCase : Union[str, Any] = { '''configuration_chinese_clip''': [ '''CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ChineseCLIPConfig''', '''ChineseCLIPOnnxConfig''', '''ChineseCLIPTextConfig''', '''ChineseCLIPVisionConfig''', ], '''processing_chinese_clip''': ['''ChineseCLIPProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Any = ['''ChineseCLIPFeatureExtractor'''] __lowerCamelCase : Optional[int] = ['''ChineseCLIPImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ '''CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ChineseCLIPModel''', '''ChineseCLIPPreTrainedModel''', '''ChineseCLIPTextModel''', '''ChineseCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys __lowerCamelCase : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
18
1
'''simple docstring''' import pytest lowercase__ : Optional[Any] = '__dummy_dataset1__' lowercase__ : str = '\nimport json\nimport os\n\nimport datasets\n\n\nREPO_URL = "https://huggingface.co/datasets/albertvillanova/tests-raw-jsonl/resolve/main/"\nURLS = {"train": REPO_URL + "wikiann-bn-train.jsonl", "validation": REPO_URL + "wikiann-bn-validation.jsonl"}\n\n\nclass __DummyDataset1__(datasets.GeneratorBasedBuilder):\n\n def _info(self):\n features = datasets.Features(\n {\n "tokens": datasets.Sequence(datasets.Value("string")),\n "ner_tags": datasets.Sequence(\n datasets.features.ClassLabel(\n names=[\n "O",\n "B-PER",\n "I-PER",\n "B-ORG",\n "I-ORG",\n "B-LOC",\n "I-LOC",\n ]\n )\n ),\n "langs": datasets.Sequence(datasets.Value("string")),\n "spans": datasets.Sequence(datasets.Value("string")),\n }\n )\n return datasets.DatasetInfo(features=features)\n\n def _split_generators(self, dl_manager):\n dl_path = dl_manager.download(URLS)\n return [\n datasets.SplitGenerator(datasets.Split.TRAIN, gen_kwargs={"filepath": dl_path["train"]}),\n datasets.SplitGenerator(datasets.Split.VALIDATION, gen_kwargs={"filepath": dl_path["validation"]}),\n ]\n\n def _generate_examples(self, filepath):\n with open(filepath, "r", encoding="utf-8") as f:\n for i, line in enumerate(f):\n yield i, json.loads(line)\n' @pytest.fixture def a__ ( ) -> int: """simple docstring""" return DATASET_LOADING_SCRIPT_NAME @pytest.fixture def a__ ( ) -> Any: """simple docstring""" return DATASET_LOADING_SCRIPT_CODE @pytest.fixture def a__ ( lowercase : Tuple, lowercase : Tuple, lowercase : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = dataset_loading_script_name _UpperCamelCase = tmp_path / '''datasets''' / script_name script_dir.mkdir(parents=lowercase ) _UpperCamelCase = script_dir / F"""{script_name}.py""" with open(lowercase, '''w''' ) as f: f.write(lowercase ) return str(lowercase )
287
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ : Optional[Any] = logging.get_logger(__name__) lowercase__ : List[Any] = { 'microsoft/biogpt': 'https://huggingface.co/microsoft/biogpt/resolve/main/config.json', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" _snake_case : Union[str, Any] = 'biogpt' def __init__( self : Optional[Any] , lowerCAmelCase__ : List[str]=42384 , lowerCAmelCase__ : Optional[int]=1024 , lowerCAmelCase__ : List[str]=24 , lowerCAmelCase__ : List[Any]=16 , lowerCAmelCase__ : Optional[int]=4096 , lowerCAmelCase__ : Optional[int]="gelu" , lowerCAmelCase__ : Optional[Any]=0.1 , lowerCAmelCase__ : Optional[int]=0.1 , lowerCAmelCase__ : Union[str, Any]=1024 , lowerCAmelCase__ : List[str]=0.02 , lowerCAmelCase__ : Tuple=1e-1_2 , lowerCAmelCase__ : Dict=True , lowerCAmelCase__ : Tuple=True , lowerCAmelCase__ : Dict=0.0 , lowerCAmelCase__ : Union[str, Any]=0.0 , lowerCAmelCase__ : Optional[int]=1 , lowerCAmelCase__ : Union[str, Any]=0 , lowerCAmelCase__ : Optional[Any]=2 , **lowerCAmelCase__ : Optional[Any] , ) -> Tuple: '''simple docstring''' _UpperCamelCase = vocab_size _UpperCamelCase = max_position_embeddings _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = scale_embedding _UpperCamelCase = use_cache _UpperCamelCase = layerdrop _UpperCamelCase = activation_dropout super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ )
287
1
"""simple docstring""" from __future__ import annotations import numpy as np def UpperCAmelCase__ ( lowerCAmelCase__ :list[float] ) -> Optional[Any]: '''simple docstring''' return np.maximum(0 , lowerCAmelCase__ ) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
197
"""simple docstring""" import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class _A : def A__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowercase = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] , mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" , up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="""text""" , addition_embed_type_num_heads=2 , cross_attention_norm="""group_norm""" , resnet_time_scale_shift="""scale_shift""" , act_fn="""gelu""" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__lowerCAmelCase , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type="""epsilon""" , variance_type="""learned_range""" , ) torch.manual_seed(0 ) lowercase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def A__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowercase = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) lowercase = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] , mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" , up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="""text""" , addition_embed_type_num_heads=2 , cross_attention_norm="""group_norm""" , resnet_time_scale_shift="""scale_shift""" , act_fn="""gelu""" , class_embed_type="""timestep""" , mid_block_scale_factor=1.4_1_4 , time_embedding_act_fn="""gelu""" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__lowerCAmelCase , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type="""epsilon""" , variance_type="""learned_range""" , ) torch.manual_seed(0 ) lowercase = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , ) torch.manual_seed(0 ) lowercase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def A__ ( self ): """simple docstring""" lowercase = self.get_dummy_components() lowercase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = inputs["""prompt"""] lowercase = inputs["""generator"""] lowercase = inputs["""num_inference_steps"""] lowercase = inputs["""output_type"""] if "image" in inputs: lowercase = inputs["""image"""] else: lowercase = None if "mask_image" in inputs: lowercase = inputs["""mask_image"""] else: lowercase = None if "original_image" in inputs: lowercase = inputs["""original_image"""] else: lowercase = None lowercase , lowercase = pipe.encode_prompt(__lowerCAmelCase ) # inputs with prompt converted to embeddings lowercase = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: lowercase = image if mask_image is not None: lowercase = mask_image if original_image is not None: lowercase = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) lowercase = pipe(**__lowerCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__lowerCAmelCase ) lowercase = self.pipeline_class.from_pretrained(__lowerCAmelCase ) pipe_loaded.to(__lowerCAmelCase ) pipe_loaded.set_progress_bar_config(disable=__lowerCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(__lowerCAmelCase , __lowerCAmelCase ) is None , f'`{optional_component}` did not stay set to None after loading.' , ) lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = inputs["""generator"""] lowercase = inputs["""num_inference_steps"""] lowercase = inputs["""output_type"""] # inputs with prompt converted to embeddings lowercase = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: lowercase = image if mask_image is not None: lowercase = mask_image if original_image is not None: lowercase = original_image lowercase = pipe_loaded(**__lowerCAmelCase )[0] lowercase = np.abs(to_np(__lowerCAmelCase ) - to_np(__lowerCAmelCase ) ).max() self.assertLess(__lowerCAmelCase , 1E-4 ) def A__ ( self ): """simple docstring""" lowercase = self.get_dummy_components() lowercase = self.pipeline_class(**__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = pipe(**__lowerCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__lowerCAmelCase ) lowercase = self.pipeline_class.from_pretrained(__lowerCAmelCase ) pipe_loaded.to(__lowerCAmelCase ) pipe_loaded.set_progress_bar_config(disable=__lowerCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase = self.get_dummy_inputs(__lowerCAmelCase ) lowercase = pipe_loaded(**__lowerCAmelCase )[0] lowercase = np.abs(to_np(__lowerCAmelCase ) - to_np(__lowerCAmelCase ) ).max() self.assertLess(__lowerCAmelCase , 1E-4 )
197
1
'''simple docstring''' from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCamelCase = logging.get_logger(__name__) # TODO Update this __lowerCamelCase = { '''facebook/esm-1b''': '''https://huggingface.co/facebook/esm-1b/resolve/main/config.json''', # See all ESM models at https://huggingface.co/models?filter=esm } class A__ ( _snake_case ): lowercase = "esm" def __init__( self , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=None , UpperCamelCase__=768 , UpperCamelCase__=12 , UpperCamelCase__=12 , UpperCamelCase__=3072 , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=1026 , UpperCamelCase__=0.02 , UpperCamelCase__=1e-1_2 , UpperCamelCase__="absolute" , UpperCamelCase__=True , UpperCamelCase__=None , UpperCamelCase__=False , UpperCamelCase__=False , UpperCamelCase__=None , UpperCamelCase__=None , **UpperCamelCase__ , ) -> Any: '''simple docstring''' super().__init__(pad_token_id=UpperCamelCase__ , mask_token_id=UpperCamelCase__ , **UpperCamelCase__ ) A_ = vocab_size A_ = hidden_size A_ = num_hidden_layers A_ = num_attention_heads A_ = intermediate_size A_ = hidden_dropout_prob A_ = attention_probs_dropout_prob A_ = max_position_embeddings A_ = initializer_range A_ = layer_norm_eps A_ = position_embedding_type A_ = use_cache A_ = emb_layer_norm_before A_ = token_dropout A_ = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("""No esmfold_config supplied for folding model, using default values.""" ) A_ = EsmFoldConfig() elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): A_ = EsmFoldConfig(**UpperCamelCase__ ) A_ = esmfold_config if vocab_list is None: logger.warning("""No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!""" ) A_ = get_default_vocab_list() else: A_ = vocab_list else: A_ = None A_ = None if self.esmfold_config is not None and getattr(self.esmfold_config , """use_esm_attn_map""" , UpperCamelCase__ ): raise ValueError("""The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!""" ) def snake_case_ ( self ) -> List[Any]: '''simple docstring''' A_ = super().to_dict() if isinstance(self.esmfold_config , UpperCamelCase__ ): A_ = self.esmfold_config.to_dict() return output @dataclass class A__ : lowercase = None lowercase = True lowercase = False lowercase = False lowercase = False lowercase = 0 lowercase = True lowercase = False lowercase = 128 lowercase = None def snake_case_ ( self ) -> Union[str, Any]: '''simple docstring''' if self.trunk is None: A_ = TrunkConfig() elif isinstance(self.trunk , UpperCamelCase__ ): A_ = TrunkConfig(**self.trunk ) def snake_case_ ( self ) -> int: '''simple docstring''' A_ = asdict(self ) A_ = self.trunk.to_dict() return output @dataclass class A__ : lowercase = 48 lowercase = 1_024 lowercase = 128 lowercase = 32 lowercase = 32 lowercase = 32 lowercase = 0 lowercase = 0 lowercase = False lowercase = 4 lowercase = 128 lowercase = None def snake_case_ ( self ) -> int: '''simple docstring''' if self.structure_module is None: A_ = StructureModuleConfig() elif isinstance(self.structure_module , UpperCamelCase__ ): A_ = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(f'''`max_recycles` should be positive, got {self.max_recycles}.''' ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( """`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got""" f''' {self.sequence_state_dim} and {self.sequence_state_dim}.''' ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( """`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got""" f''' {self.pairwise_state_dim} and {self.pairwise_state_dim}.''' ) A_ = self.sequence_state_dim // self.sequence_head_width A_ = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( """`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got""" f''' {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}.''' ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( """`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got""" f''' {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}.''' ) if self.pairwise_state_dim % 2 != 0: raise ValueError(f'''`pairwise_state_dim` should be even, got {self.pairwise_state_dim}.''' ) if self.dropout >= 0.4: raise ValueError(f'''`dropout` should not be greater than 0.4, got {self.dropout}.''' ) def snake_case_ ( self ) -> int: '''simple docstring''' A_ = asdict(self ) A_ = self.structure_module.to_dict() return output @dataclass class A__ : lowercase = 384 lowercase = 128 lowercase = 16 lowercase = 128 lowercase = 12 lowercase = 4 lowercase = 8 lowercase = 0.1 lowercase = 8 lowercase = 1 lowercase = 2 lowercase = 7 lowercase = 10 lowercase = 1e-8 lowercase = 1e5 def snake_case_ ( self ) -> List[str]: '''simple docstring''' return asdict(self ) def UpperCAmelCase__ ( ) -> Dict: return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
366
'''simple docstring''' import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa __lowerCamelCase = logging.getLogger(__name__) class A__ ( _snake_case ): lowercase = "summarization" lowercase = ["loss"] lowercase = ROUGE_KEYS lowercase = "rouge2" def __init__( self , UpperCamelCase__ , **UpperCamelCase__ ) -> Dict: '''simple docstring''' if hparams.sortish_sampler and hparams.gpus > 1: A_ = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError("""Dynamic Batch size does not work for multi-gpu training""" ) if hparams.sortish_sampler: raise ValueError("""--sortish_sampler and --max_tokens_per_batch may not be used simultaneously""" ) super().__init__(UpperCamelCase__ , num_labels=UpperCamelCase__ , mode=self.mode , **UpperCamelCase__ ) use_task_specific_params(self.model , """summarization""" ) save_git_info(self.hparams.output_dir ) A_ = Path(self.output_dir ) / """metrics.json""" A_ = Path(self.output_dir ) / """hparams.pkl""" pickle_save(self.hparams , self.hparams_save_path ) A_ = 0 A_ = defaultdict(UpperCamelCase__ ) A_ = self.config.model_type A_ = self.config.tgt_vocab_size if self.model_type == """fsmt""" else self.config.vocab_size A_ = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } A_ = { """train""": self.hparams.n_train, """val""": self.hparams.n_val, """test""": self.hparams.n_test, } A_ = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} A_ = { """train""": self.hparams.max_target_length, """val""": self.hparams.val_max_target_length, """test""": self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], f'''target_lens: {self.target_lens}''' assert self.target_lens["train"] <= self.target_lens["test"], f'''target_lens: {self.target_lens}''' if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) A_ = get_git_info()["""repo_sha"""] A_ = hparams.num_workers A_ = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , UpperCamelCase__ ): A_ = self.tokenizer.lang_code_to_id[hparams.tgt_lang] A_ = self.decoder_start_token_id A_ = ( SeqaSeqDataset if hasattr(self.tokenizer , """prepare_seq2seq_batch""" ) else LegacySeqaSeqDataset ) A_ = False A_ = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: A_ = self.hparams.eval_max_gen_length else: A_ = self.model.config.max_length A_ = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def snake_case_ ( self , UpperCamelCase__ ) -> Dict[str, List[str]]: '''simple docstring''' A_ = { k: self.tokenizer.batch_decode(v.tolist() ) if """mask""" not in k else v.shape for k, v in batch.items() } save_json(UpperCamelCase__ , Path(self.output_dir ) / """text_batch.json""" ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / """tok_batch.json""" ) A_ = True return readable_batch def snake_case_ ( self , UpperCamelCase__ , **UpperCamelCase__ ) -> str: '''simple docstring''' return self.model(UpperCamelCase__ , **UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ ) -> List[str]: '''simple docstring''' A_ = self.tokenizer.batch_decode( UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ ) return lmap(str.strip , UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ ) -> Tuple: '''simple docstring''' A_ = self.tokenizer.pad_token_id A_ , A_ = batch["""input_ids"""], batch["""attention_mask"""] A_ = batch["""labels"""] if isinstance(self.model , UpperCamelCase__ ): A_ = self.model._shift_right(UpperCamelCase__ ) else: A_ = shift_tokens_right(UpperCamelCase__ , UpperCamelCase__ ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero A_ = decoder_input_ids self.save_readable_batch(UpperCamelCase__ ) A_ = self(UpperCamelCase__ , attention_mask=UpperCamelCase__ , decoder_input_ids=UpperCamelCase__ , use_cache=UpperCamelCase__ ) A_ = outputs["""logits"""] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id A_ = nn.CrossEntropyLoss(ignore_index=UpperCamelCase__ ) assert lm_logits.shape[-1] == self.vocab_size A_ = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: A_ = nn.functional.log_softmax(UpperCamelCase__ , dim=-1 ) A_ , A_ = label_smoothed_nll_loss( UpperCamelCase__ , UpperCamelCase__ , self.hparams.label_smoothing , ignore_index=UpperCamelCase__ ) return (loss,) @property def snake_case_ ( self ) -> int: '''simple docstring''' return self.tokenizer.pad_token_id def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' A_ = self._step(UpperCamelCase__ ) A_ = dict(zip(self.loss_names , UpperCamelCase__ ) ) # tokens per batch A_ = batch["""input_ids"""].ne(self.pad ).sum() + batch["""labels"""].ne(self.pad ).sum() A_ = batch["""input_ids"""].shape[0] A_ = batch["""input_ids"""].eq(self.pad ).sum() A_ = batch["""input_ids"""].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' return self._generative_step(UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__="val" ) -> Dict: '''simple docstring''' self.step_count += 1 A_ = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} A_ = losses["""loss"""] A_ = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ["""gen_time""", """gen_len"""] } A_ = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) A_ = torch.tensor(UpperCamelCase__ ).type_as(UpperCamelCase__ ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(UpperCamelCase__ ) A_ = {f'''{prefix}_avg_{k}''': x for k, x in losses.items()} A_ = self.step_count self.metrics[prefix].append(UpperCamelCase__ ) # callback writes this to self.metrics_save_path A_ = flatten_list([x["""preds"""] for x in outputs] ) return { "log": all_metrics, "preds": preds, f'''{prefix}_loss''': loss, f'''{prefix}_{self.val_metric}''': metric_tensor, } def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' return calculate_rouge(UpperCamelCase__ , UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ ) -> dict: '''simple docstring''' A_ = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') A_ = self.model.generate( batch["""input_ids"""] , attention_mask=batch["""attention_mask"""] , use_cache=UpperCamelCase__ , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) A_ = (time.time() - ta) / batch["""input_ids"""].shape[0] A_ = self.ids_to_clean_text(UpperCamelCase__ ) A_ = self.ids_to_clean_text(batch["""labels"""] ) A_ = self._step(UpperCamelCase__ ) A_ = dict(zip(self.loss_names , UpperCamelCase__ ) ) A_ = self.calc_generative_metrics(UpperCamelCase__ , UpperCamelCase__ ) A_ = np.mean(lmap(UpperCamelCase__ , UpperCamelCase__ ) ) base_metrics.update(gen_time=UpperCamelCase__ , gen_len=UpperCamelCase__ , preds=UpperCamelCase__ , target=UpperCamelCase__ , **UpperCamelCase__ ) return base_metrics def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ ) -> Tuple: '''simple docstring''' return self._generative_step(UpperCamelCase__ ) def snake_case_ ( self , UpperCamelCase__ ) -> Dict: '''simple docstring''' return self.validation_epoch_end(UpperCamelCase__ , prefix="""test""" ) def snake_case_ ( self , UpperCamelCase__ ) -> SeqaSeqDataset: '''simple docstring''' A_ = self.n_obs[type_path] A_ = self.target_lens[type_path] A_ = self.dataset_class( self.tokenizer , type_path=UpperCamelCase__ , n_obs=UpperCamelCase__ , max_target_length=UpperCamelCase__ , **self.dataset_kwargs , ) return dataset def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = False ) -> DataLoader: '''simple docstring''' A_ = self.get_dataset(UpperCamelCase__ ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": A_ = dataset.make_sortish_sampler(UpperCamelCase__ , distributed=self.hparams.gpus > 1 ) return DataLoader( UpperCamelCase__ , batch_size=UpperCamelCase__ , collate_fn=dataset.collate_fn , shuffle=UpperCamelCase__ , num_workers=self.num_workers , sampler=UpperCamelCase__ , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": A_ = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( UpperCamelCase__ , batch_sampler=UpperCamelCase__ , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( UpperCamelCase__ , batch_size=UpperCamelCase__ , collate_fn=dataset.collate_fn , shuffle=UpperCamelCase__ , num_workers=self.num_workers , sampler=UpperCamelCase__ , ) def snake_case_ ( self ) -> DataLoader: '''simple docstring''' A_ = self.get_dataloader("""train""" , batch_size=self.hparams.train_batch_size , shuffle=UpperCamelCase__ ) return dataloader def snake_case_ ( self ) -> DataLoader: '''simple docstring''' return self.get_dataloader("""val""" , batch_size=self.hparams.eval_batch_size ) def snake_case_ ( self ) -> DataLoader: '''simple docstring''' return self.get_dataloader("""test""" , batch_size=self.hparams.eval_batch_size ) @staticmethod def snake_case_ ( UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' BaseTransformer.add_model_specific_args(UpperCamelCase__ , UpperCamelCase__ ) add_generic_args(UpperCamelCase__ , UpperCamelCase__ ) parser.add_argument( """--max_source_length""" , default=1024 , type=UpperCamelCase__ , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--max_target_length""" , default=56 , type=UpperCamelCase__ , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--val_max_target_length""" , default=142 , type=UpperCamelCase__ , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--test_max_target_length""" , default=142 , type=UpperCamelCase__ , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument("""--freeze_encoder""" , action="""store_true""" ) parser.add_argument("""--freeze_embeds""" , action="""store_true""" ) parser.add_argument("""--sortish_sampler""" , action="""store_true""" , default=UpperCamelCase__ ) parser.add_argument("""--overwrite_output_dir""" , action="""store_true""" , default=UpperCamelCase__ ) parser.add_argument("""--max_tokens_per_batch""" , type=UpperCamelCase__ , default=UpperCamelCase__ ) parser.add_argument("""--logger_name""" , type=UpperCamelCase__ , choices=["""default""", """wandb""", """wandb_shared"""] , default="""default""" ) parser.add_argument("""--n_train""" , type=UpperCamelCase__ , default=-1 , required=UpperCamelCase__ , help="""# examples. -1 means use all.""" ) parser.add_argument("""--n_val""" , type=UpperCamelCase__ , default=500 , required=UpperCamelCase__ , help="""# examples. -1 means use all.""" ) parser.add_argument("""--n_test""" , type=UpperCamelCase__ , default=-1 , required=UpperCamelCase__ , help="""# examples. -1 means use all.""" ) parser.add_argument( """--task""" , type=UpperCamelCase__ , default="""summarization""" , required=UpperCamelCase__ , help="""# examples. -1 means use all.""" ) parser.add_argument("""--label_smoothing""" , type=UpperCamelCase__ , default=0.0 , required=UpperCamelCase__ ) parser.add_argument("""--src_lang""" , type=UpperCamelCase__ , default="""""" , required=UpperCamelCase__ ) parser.add_argument("""--tgt_lang""" , type=UpperCamelCase__ , default="""""" , required=UpperCamelCase__ ) parser.add_argument("""--eval_beams""" , type=UpperCamelCase__ , default=UpperCamelCase__ , required=UpperCamelCase__ ) parser.add_argument( """--val_metric""" , type=UpperCamelCase__ , default=UpperCamelCase__ , required=UpperCamelCase__ , choices=["""bleu""", """rouge2""", """loss""", None] ) parser.add_argument("""--eval_max_gen_length""" , type=UpperCamelCase__ , default=UpperCamelCase__ , help="""never generate more than n tokens""" ) parser.add_argument("""--save_top_k""" , type=UpperCamelCase__ , default=1 , required=UpperCamelCase__ , help="""How many checkpoints to save""" ) parser.add_argument( """--early_stopping_patience""" , type=UpperCamelCase__ , default=-1 , required=UpperCamelCase__ , help=( """-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So""" """ val_check_interval will effect it.""" ) , ) return parser class A__ ( _snake_case ): lowercase = "translation" lowercase = ["loss"] lowercase = ["bleu"] lowercase = "bleu" def __init__( self , UpperCamelCase__ , **UpperCamelCase__ ) -> Optional[int]: '''simple docstring''' super().__init__(UpperCamelCase__ , **UpperCamelCase__ ) A_ = hparams.src_lang A_ = hparams.tgt_lang def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ ) -> dict: '''simple docstring''' return calculate_bleu(UpperCamelCase__ , UpperCamelCase__ ) def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__=None ) -> SummarizationModule: Path(args.output_dir ).mkdir(exist_ok=UpperCAmelCase__ ) check_output_dir(UpperCAmelCase__, expected_items=3 ) if model is None: if "summarization" in args.task: A_ = SummarizationModule(UpperCAmelCase__ ) else: A_ = TranslationModule(UpperCAmelCase__ ) A_ = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith("""/tmp""" ) or str(args.output_dir ).startswith("""/var""" ) ): A_ = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger A_ = os.environ.get("""WANDB_PROJECT""", UpperCAmelCase__ ) A_ = WandbLogger(name=model.output_dir.name, project=UpperCAmelCase__ ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger A_ = WandbLogger(name=model.output_dir.name, project=F'''hf_{dataset}''' ) if args.early_stopping_patience >= 0: A_ = get_early_stopping_callback(model.val_metric, args.early_stopping_patience ) else: A_ = False A_ = args.val_metric == """loss""" A_ = generic_train( UpperCAmelCase__, UpperCAmelCase__, logging_callback=SeqaSeqLoggingCallback(), checkpoint_callback=get_checkpoint_callback( args.output_dir, model.val_metric, args.save_top_k, UpperCAmelCase__ ), early_stopping_callback=UpperCAmelCase__, logger=UpperCAmelCase__, ) pickle_save(model.hparams, model.output_dir / """hparams.pkl""" ) if not args.do_predict: return model A_ = """""" A_ = sorted(glob.glob(os.path.join(args.output_dir, """*.ckpt""" ), recursive=UpperCAmelCase__ ) ) if checkpoints: A_ = checkpoints[-1] A_ = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": __lowerCamelCase = argparse.ArgumentParser() __lowerCamelCase = pl.Trainer.add_argparse_args(parser) __lowerCamelCase = SummarizationModule.add_model_specific_args(parser, os.getcwd()) __lowerCamelCase = parser.parse_args() main(args)
101
0
import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel _A = { '''gwf-440k''': { '''url''': '''https://model-server.zqevans2.workers.dev/gwf-440k.ckpt''', '''sample_rate''': 48_000, '''sample_size''': 65_536, }, '''jmann-small-190k''': { '''url''': '''https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt''', '''sample_rate''': 48_000, '''sample_size''': 65_536, }, '''jmann-large-580k''': { '''url''': '''https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt''', '''sample_rate''': 48_000, '''sample_size''': 131_072, }, '''maestro-uncond-150k''': { '''url''': '''https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt''', '''sample_rate''': 16_000, '''sample_size''': 65_536, }, '''unlocked-uncond-250k''': { '''url''': '''https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt''', '''sample_rate''': 16_000, '''sample_size''': 65_536, }, '''honk-140k''': { '''url''': '''https://model-server.zqevans2.workers.dev/honk-140k.ckpt''', '''sample_rate''': 16_000, '''sample_size''': 65_536, }, } def lowerCamelCase__ ( a__ : Optional[Any] , a__ : Optional[Any] ) -> List[str]: return torch.atana(a__ , a__ ) / math.pi * 2 def lowerCamelCase__ ( a__ : Tuple ) -> str: UpperCamelCase_ = torch.sin(t * math.pi / 2 ) ** 2 UpperCamelCase_ = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(a__ , a__ ) class lowercase_ ( __SCREAMING_SNAKE_CASE ): pass class lowercase_ ( nn.Module ): def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__() UpperCamelCase_ = DiffusionAttnUnetaD(__UpperCamelCase , n_attn_layers=4 ) UpperCamelCase_ = deepcopy(self.diffusion ) UpperCamelCase_ = torch.quasirandom.SobolEngine(1 , scramble=__UpperCamelCase ) def lowerCamelCase__ ( a__ : Union[str, Any] ) -> Dict: UpperCamelCase_ = MODELS_MAP[model_name]["""url"""] os.system(f'''wget {url} ./''' ) return f'''./{model_name}.ckpt''' _A = { '''1''': '''resnets.0''', '''2''': '''attentions.0''', '''3''': '''resnets.1''', '''4''': '''attentions.1''', '''5''': '''resnets.2''', '''6''': '''attentions.2''', } _A = { '''8''': '''resnets.0''', '''9''': '''attentions.0''', '''10''': '''resnets.1''', '''11''': '''attentions.1''', '''12''': '''resnets.2''', '''13''': '''attentions.2''', } _A = { '''1''': '''resnets.0''', '''2''': '''attentions.0''', '''3''': '''resnets.1''', '''4''': '''attentions.1''', '''5''': '''resnets.2''', '''6''': '''attentions.2''', '''8''': '''resnets.3''', '''9''': '''attentions.3''', '''10''': '''resnets.4''', '''11''': '''attentions.4''', '''12''': '''resnets.5''', '''13''': '''attentions.5''', } _A = { '''0''': '''resnets.0''', '''1''': '''resnets.1''', '''2''': '''resnets.2''', '''4''': '''resnets.0''', '''5''': '''resnets.1''', '''6''': '''resnets.2''', } _A = { '''skip''': '''conv_skip''', '''main.0''': '''conv_1''', '''main.1''': '''group_norm_1''', '''main.3''': '''conv_2''', '''main.4''': '''group_norm_2''', } _A = { '''norm''': '''group_norm''', '''qkv_proj''': ['''query''', '''key''', '''value'''], '''out_proj''': ['''proj_attn'''], } def lowerCamelCase__ ( a__ : List[str] ) -> List[Any]: if name.startswith("""skip""" ): return name.replace("""skip""" , RES_CONV_MAP["""skip"""] ) # name has to be of format main.{digit} if not name.startswith("""main.""" ): raise ValueError(f'''ResConvBlock error with {name}''' ) return name.replace(name[:6] , RES_CONV_MAP[name[:6]] ) def lowerCamelCase__ ( a__ : int ) -> Optional[int]: for key, value in ATTN_MAP.items(): if name.startswith(a__ ) and not isinstance(a__ , a__ ): return name.replace(a__ , a__ ) elif name.startswith(a__ ): return [name.replace(a__ , a__ ) for v in value] raise ValueError(f'''Attn error with {name}''' ) def lowerCamelCase__ ( a__ : Tuple , a__ : Dict=13 ) -> str: UpperCamelCase_ = input_string if string.split(""".""" )[0] == "timestep_embed": return string.replace("""timestep_embed""" , """time_proj""" ) UpperCamelCase_ = 0 if string.startswith("""net.3.""" ): depth += 1 UpperCamelCase_ = string[6:] elif string.startswith("""net.""" ): UpperCamelCase_ = string[4:] while string.startswith("""main.7.""" ): depth += 1 UpperCamelCase_ = string[7:] if string.startswith("""main.""" ): UpperCamelCase_ = string[5:] # mid block if string[:2].isdigit(): UpperCamelCase_ = string[:2] UpperCamelCase_ = string[2:] else: UpperCamelCase_ = string[0] UpperCamelCase_ = string[1:] if depth == max_depth: UpperCamelCase_ = MID_NUM_TO_LAYER[layer_num] UpperCamelCase_ = """mid_block""" elif depth > 0 and int(a__ ) < 7: UpperCamelCase_ = DOWN_NUM_TO_LAYER[layer_num] UpperCamelCase_ = f'''down_blocks.{depth}''' elif depth > 0 and int(a__ ) > 7: UpperCamelCase_ = UP_NUM_TO_LAYER[layer_num] UpperCamelCase_ = f'''up_blocks.{max_depth - depth - 1}''' elif depth == 0: UpperCamelCase_ = DEPTH_0_TO_LAYER[layer_num] UpperCamelCase_ = f'''up_blocks.{max_depth - 1}''' if int(a__ ) > 3 else """down_blocks.0""" if not string_left.startswith(""".""" ): raise ValueError(f'''Naming error with {input_string} and string_left: {string_left}.''' ) UpperCamelCase_ = string_left[1:] if "resnets" in new_layer: UpperCamelCase_ = convert_resconv_naming(a__ ) elif "attentions" in new_layer: UpperCamelCase_ = convert_attn_naming(a__ ) UpperCamelCase_ = new_string_left if not isinstance(a__ , a__ ): UpperCamelCase_ = prefix + """.""" + new_layer + """.""" + string_left else: UpperCamelCase_ = [prefix + """.""" + new_layer + """.""" + s for s in string_left] return new_string def lowerCamelCase__ ( a__ : Dict ) -> List[Any]: UpperCamelCase_ = {} for k, v in state_dict.items(): if k.endswith("""kernel""" ): # up- and downsample layers, don't have trainable weights continue UpperCamelCase_ = rename(a__ ) # check if we need to transform from Conv => Linear for attention if isinstance(a__ , a__ ): UpperCamelCase_ = transform_conv_attns(a__ , a__ , a__ ) else: UpperCamelCase_ = v return new_state_dict def lowerCamelCase__ ( a__ : Dict , a__ : Optional[Any] , a__ : Optional[Any] ) -> List[Any]: if len(a__ ) == 1: if len(v.shape ) == 3: # weight UpperCamelCase_ = v[:, :, 0] else: # bias UpperCamelCase_ = v else: # qkv matrices UpperCamelCase_ = v.shape[0] UpperCamelCase_ = trippled_shape // 3 for i in range(3 ): if len(v.shape ) == 3: UpperCamelCase_ = v[i * single_shape : (i + 1) * single_shape, :, 0] else: UpperCamelCase_ = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def lowerCamelCase__ ( a__ : Any ) -> Optional[int]: UpperCamelCase_ = torch.device("""cuda""" if torch.cuda.is_available() else """cpu""" ) UpperCamelCase_ = args.model_path.split("""/""" )[-1].split(""".""" )[0] if not os.path.isfile(args.model_path ): assert ( model_name == args.model_path ), f'''Make sure to provide one of the official model names {MODELS_MAP.keys()}''' UpperCamelCase_ = download(a__ ) UpperCamelCase_ = MODELS_MAP[model_name]["""sample_rate"""] UpperCamelCase_ = MODELS_MAP[model_name]["""sample_size"""] UpperCamelCase_ = Object() UpperCamelCase_ = sample_size UpperCamelCase_ = sample_rate UpperCamelCase_ = 0 UpperCamelCase_ = UNetaDModel(sample_size=a__ , sample_rate=a__ ) UpperCamelCase_ = diffusers_model.state_dict() UpperCamelCase_ = DiffusionUncond(a__ ) orig_model.load_state_dict(torch.load(args.model_path , map_location=a__ )["""state_dict"""] ) UpperCamelCase_ = orig_model.diffusion_ema.eval() UpperCamelCase_ = orig_model.state_dict() UpperCamelCase_ = rename_orig_weights(a__ ) UpperCamelCase_ = set(renamed_state_dict.keys() ) - set(diffusers_state_dict.keys() ) UpperCamelCase_ = set(diffusers_state_dict.keys() ) - set(renamed_state_dict.keys() ) assert len(a__ ) == 0, f'''Problem with {renamed_minus_diffusers}''' assert all(k.endswith("""kernel""" ) for k in list(a__ ) ), f'''Problem with {diffusers_minus_renamed}''' for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), f'''Shape for {key} doesn\'t match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}''' if key == "time_proj.weight": UpperCamelCase_ = value.squeeze() UpperCamelCase_ = value diffusers_model.load_state_dict(a__ ) UpperCamelCase_ = 100 UpperCamelCase_ = 33 UpperCamelCase_ = IPNDMScheduler(num_train_timesteps=a__ ) UpperCamelCase_ = torch.manual_seed(a__ ) UpperCamelCase_ = torch.randn([1, 2, config.sample_size] , generator=a__ ).to(a__ ) UpperCamelCase_ = torch.linspace(1 , 0 , steps + 1 , device=a__ )[:-1] UpperCamelCase_ = get_crash_schedule(a__ ) UpperCamelCase_ = DanceDiffusionPipeline(unet=a__ , scheduler=a__ ) UpperCamelCase_ = torch.manual_seed(33 ) UpperCamelCase_ = pipe(num_inference_steps=a__ , generator=a__ ).audios UpperCamelCase_ = sampling.iplms_sample(a__ , a__ , a__ , {} ) UpperCamelCase_ = generated.clamp(-1 , 1 ) UpperCamelCase_ = (generated - audio).abs().sum() UpperCamelCase_ = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path ) print("""Diff sum""" , a__ ) print("""Diff max""" , a__ ) assert diff_max < 1e-3, f'''Diff max: {diff_max} is too much :-/''' print(f'''Conversion for {model_name} successful!''' ) if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument('''--model_path''', default=None, type=str, required=True, help='''Path to the model to convert.''') parser.add_argument( '''--save''', default=True, type=bool, required=False, help='''Whether to save the converted model or not.''' ) parser.add_argument('''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the output model.''') _A = parser.parse_args() main(args)
122
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class lowercase_ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): A__ : Dict = BarthezTokenizer A__ : List[Any] = BarthezTokenizerFast A__ : int = True A__ : str = True def lowerCamelCase_ ( self ): """simple docstring""" super().setUp() UpperCamelCase_ = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=__UpperCamelCase ) UpperCamelCase_ = tokenizer def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = """<pad>""" UpperCamelCase_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCamelCase ) , __UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCamelCase ) , __UpperCamelCase ) def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(__UpperCamelCase ) , 1_0_1_1_2_2 ) def lowerCamelCase_ ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2 ) @require_torch def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] UpperCamelCase_ = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] UpperCamelCase_ = self.tokenizer( __UpperCamelCase , max_length=len(__UpperCamelCase ) , padding=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors="""pt""" ) self.assertIsInstance(__UpperCamelCase , __UpperCamelCase ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) UpperCamelCase_ = batch.input_ids.tolist()[0] self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) def lowerCamelCase_ ( self ): """simple docstring""" if not self.test_rust_tokenizer: return UpperCamelCase_ = self.get_tokenizer() UpperCamelCase_ = self.get_rust_tokenizer() UpperCamelCase_ = """I was born in 92000, and this is falsé.""" UpperCamelCase_ = tokenizer.tokenize(__UpperCamelCase ) UpperCamelCase_ = rust_tokenizer.tokenize(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) UpperCamelCase_ = tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) UpperCamelCase_ = rust_tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) UpperCamelCase_ = self.get_rust_tokenizer() UpperCamelCase_ = tokenizer.encode(__UpperCamelCase ) UpperCamelCase_ = rust_tokenizer.encode(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) @slow def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = {"""input_ids""": [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. UpperCamelCase_ = [ """Le transformeur est un modèle d'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=__UpperCamelCase , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=__UpperCamelCase , )
122
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer __A =logging.get_logger(__name__) __A ={'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __A ={ '''vocab_file''': { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt''' ), '''google/realm-orqa-nq-openqa''': '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt''', '''google/realm-orqa-nq-reader''': '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt''', '''google/realm-orqa-wq-openqa''': '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt''', '''google/realm-orqa-wq-reader''': '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt''', }, '''tokenizer_file''': { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json''' ), '''google/realm-orqa-nq-openqa''': ( '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json''' ), '''google/realm-orqa-nq-reader''': ( '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json''' ), '''google/realm-orqa-wq-openqa''': ( '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json''' ), '''google/realm-orqa-wq-reader''': ( '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json''' ), }, } __A ={ '''google/realm-cc-news-pretrained-embedder''': 5_1_2, '''google/realm-cc-news-pretrained-encoder''': 5_1_2, '''google/realm-cc-news-pretrained-scorer''': 5_1_2, '''google/realm-cc-news-pretrained-openqa''': 5_1_2, '''google/realm-orqa-nq-openqa''': 5_1_2, '''google/realm-orqa-nq-reader''': 5_1_2, '''google/realm-orqa-wq-openqa''': 5_1_2, '''google/realm-orqa-wq-reader''': 5_1_2, } __A ={ '''google/realm-cc-news-pretrained-embedder''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-encoder''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-scorer''': {'''do_lower_case''': True}, '''google/realm-cc-news-pretrained-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-nq-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-nq-reader''': {'''do_lower_case''': True}, '''google/realm-orqa-wq-openqa''': {'''do_lower_case''': True}, '''google/realm-orqa-wq-reader''': {'''do_lower_case''': True}, } class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_INIT_CONFIGURATION lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = RealmTokenizer def __init__( self , lowercase=None , lowercase=None , lowercase=True , lowercase="[UNK]" , lowercase="[SEP]" , lowercase="[PAD]" , lowercase="[CLS]" , lowercase="[MASK]" , lowercase=True , lowercase=None , **lowercase , ) -> Optional[int]: super().__init__( lowercase , tokenizer_file=lowercase , do_lower_case=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , tokenize_chinese_chars=lowercase , strip_accents=lowercase , **lowercase , ) lowerCamelCase_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , lowercase ) != do_lower_case or normalizer_state.get("strip_accents" , lowercase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , lowercase ) != tokenize_chinese_chars ): lowerCamelCase_ = getattr(lowercase , normalizer_state.pop("type" ) ) lowerCamelCase_ = do_lower_case lowerCamelCase_ = strip_accents lowerCamelCase_ = tokenize_chinese_chars lowerCamelCase_ = normalizer_class(**lowercase ) lowerCamelCase_ = do_lower_case def SCREAMING_SNAKE_CASE_( self , lowercase , **lowercase ) -> Optional[Any]: lowerCamelCase_ = PaddingStrategy.MAX_LENGTH lowerCamelCase_ = text lowerCamelCase_ = kwargs.pop("text_pair" , lowercase ) lowerCamelCase_ = kwargs.pop("return_tensors" , lowercase ) lowerCamelCase_ = { "input_ids": [], "attention_mask": [], "token_type_ids": [], } for idx, candidate_text in enumerate(lowercase ): if batch_text_pair is not None: lowerCamelCase_ = batch_text_pair[idx] else: lowerCamelCase_ = None lowerCamelCase_ = super().__call__(lowercase , lowercase , return_tensors=lowercase , **lowercase ) lowerCamelCase_ = encoded_candidates.get("input_ids" ) lowerCamelCase_ = encoded_candidates.get("attention_mask" ) lowerCamelCase_ = encoded_candidates.get("token_type_ids" ) if encoded_input_ids is not None: output_data["input_ids"].append(lowercase ) if encoded_attention_mask is not None: output_data["attention_mask"].append(lowercase ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(lowercase ) lowerCamelCase_ = {key: item for key, item in output_data.items() if len(lowercase ) != 0} return BatchEncoding(lowercase , tensor_type=lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase=None ) -> int: lowerCamelCase_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> Tuple[str]: lowerCamelCase_ = self._tokenizer.model.save(lowercase , name=lowercase ) return tuple(lowercase )
369
import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class _SCREAMING_SNAKE_CASE : def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=False , lowercase=True , lowercase=99 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.0_2 , lowercase=3 , lowercase=4 , lowercase=None , ) -> Optional[int]: lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = seq_length lowerCamelCase_ = is_training lowerCamelCase_ = use_input_mask lowerCamelCase_ = use_token_type_ids lowerCamelCase_ = use_labels lowerCamelCase_ = vocab_size lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = max_position_embeddings lowerCamelCase_ = type_vocab_size lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = num_labels lowerCamelCase_ = num_choices lowerCamelCase_ = scope def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCamelCase_ = None if self.use_input_mask: lowerCamelCase_ = random_attention_mask([self.batch_size, self.seq_length] ) lowerCamelCase_ = None if self.use_token_type_ids: lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCamelCase_ = None lowerCamelCase_ = None lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCamelCase_ = ids_tensor([self.batch_size] , self.num_choices ) lowerCamelCase_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE_( self ) -> Tuple: return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Union[str, Any]: lowerCamelCase_ = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase ) lowerCamelCase_ = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> Any: lowerCamelCase_ = BioGptForCausalLM(config=lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ) -> Tuple: lowerCamelCase_ = BioGptModel(config=lowercase ) model.to(lowercase ) model.eval() # create attention mask lowerCamelCase_ = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) lowerCamelCase_ = self.seq_length // 2 lowerCamelCase_ = 0 # first forward pass lowerCamelCase_ , lowerCamelCase_ = model(lowercase , attention_mask=lowercase ).to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase_ = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids lowerCamelCase_ = ids_tensor((1,) , lowercase ).item() + 1 lowerCamelCase_ = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) lowerCamelCase_ = random_other_next_tokens # append to next input_ids and attn_mask lowerCamelCase_ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase_ = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=lowercase )] , dim=1 , ) # get two different outputs lowerCamelCase_ = model(lowercase , attention_mask=lowercase )["last_hidden_state"] lowerCamelCase_ = model(lowercase , past_key_values=lowercase , attention_mask=lowercase )["last_hidden_state"] # select random slice lowerCamelCase_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase_ = output_from_no_past[:, -1, random_slice_idx].detach() lowerCamelCase_ = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1e-3 ) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ) -> Dict: lowerCamelCase_ = BioGptModel(config=lowercase ).to(lowercase ).eval() lowerCamelCase_ = torch.ones(input_ids.shape , dtype=torch.long , device=lowercase ) # first forward pass lowerCamelCase_ = model(lowercase , attention_mask=lowercase , use_cache=lowercase ) lowerCamelCase_ , lowerCamelCase_ = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids lowerCamelCase_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowerCamelCase_ = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and lowerCamelCase_ = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCamelCase_ = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) lowerCamelCase_ = model(lowercase , attention_mask=lowercase )["last_hidden_state"] lowerCamelCase_ = model(lowercase , attention_mask=lowercase , past_key_values=lowercase )[ "last_hidden_state" ] # select random slice lowerCamelCase_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCamelCase_ = output_from_no_past[:, -3:, random_slice_idx].detach() lowerCamelCase_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase , lowercase , atol=1e-3 ) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase , lowercase=False ) -> Optional[int]: lowerCamelCase_ = BioGptForCausalLM(lowercase ) model.to(lowercase ) if gradient_checkpointing: model.gradient_checkpointing_enable() lowerCamelCase_ = model(lowercase , labels=lowercase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def SCREAMING_SNAKE_CASE_( self , lowercase , *lowercase ) -> List[Any]: lowerCamelCase_ = BioGptModel(lowercase ) lowerCamelCase_ = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.0_0_1 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.0_1 ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase , *lowercase ) -> Any: lowerCamelCase_ = self.num_labels lowerCamelCase_ = BioGptForTokenClassification(lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = self.prepare_config_and_inputs() ( ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ( lowerCamelCase_ ) , ) = config_and_inputs lowerCamelCase_ = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): lowerCAmelCase__ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) lowerCAmelCase__ = (BioGptForCausalLM,) if is_torch_available() else () lowerCAmelCase__ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) lowerCAmelCase__ = False def SCREAMING_SNAKE_CASE_( self ) -> Any: lowerCamelCase_ = BioGptModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCamelCase_ = type self.model_tester.create_and_check_model(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*lowercase , gradient_checkpointing=lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*lowercase ) @slow def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) model.to(lowercase ) lowerCamelCase_ = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) lowerCamelCase_ = "left" # Define PAD Token = EOS Token = 50256 lowerCamelCase_ = tokenizer.eos_token lowerCamelCase_ = model.config.eos_token_id # use different length sentences to test batching lowerCamelCase_ = [ "Hello, my dog is a little", "Today, I", ] lowerCamelCase_ = tokenizer(lowercase , return_tensors="pt" , padding=lowercase ) lowerCamelCase_ = inputs["input_ids"].to(lowercase ) lowerCamelCase_ = model.generate( input_ids=lowercase , attention_mask=inputs["attention_mask"].to(lowercase ) , ) lowerCamelCase_ = tokenizer(sentences[0] , return_tensors="pt" ).input_ids.to(lowercase ) lowerCamelCase_ = model.generate(input_ids=lowercase ) lowerCamelCase_ = inputs_non_padded.shape[-1] - inputs["attention_mask"][-1].long().sum().cpu().item() lowerCamelCase_ = tokenizer(sentences[1] , return_tensors="pt" ).input_ids.to(lowercase ) lowerCamelCase_ = model.generate(input_ids=lowercase , max_length=model.config.max_length - num_paddings ) lowerCamelCase_ = tokenizer.batch_decode(lowercase , skip_special_tokens=lowercase ) lowerCamelCase_ = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowercase ) lowerCamelCase_ = tokenizer.decode(output_padded[0] , skip_special_tokens=lowercase ) lowerCamelCase_ = [ "Hello, my dog is a little bit bigger than a little bit.", "Today, I have a good idea of how to use the information", ] self.assertListEqual(lowercase , lowercase ) self.assertListEqual(lowercase , [non_padded_sentence, padded_sentence] ) @slow def SCREAMING_SNAKE_CASE_( self ) -> int: for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ = BioGptModel.from_pretrained(lowercase ) self.assertIsNotNone(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Tuple: lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase_ = 3 lowerCamelCase_ = input_dict["input_ids"] lowerCamelCase_ = input_ids.ne(1 ).to(lowercase ) lowerCamelCase_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowerCamelCase_ = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase_ = 3 lowerCamelCase_ = "multi_label_classification" lowerCamelCase_ = input_dict["input_ids"] lowerCamelCase_ = input_ids.ne(1 ).to(lowercase ) lowerCamelCase_ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) lowerCamelCase_ = BioGptForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() lowerCamelCase_ = model(lowercase , attention_mask=lowercase , labels=lowercase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) lowerCamelCase_ = torch.tensor([[2, 4805, 9, 656, 21]] ) lowerCamelCase_ = model(lowercase )[0] lowerCamelCase_ = 42384 lowerCamelCase_ = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , lowercase ) lowerCamelCase_ = torch.tensor( [[[-9.5_2_3_6, -9.8_9_1_8, 1_0.4_5_5_7], [-1_1.0_4_6_9, -9.6_4_2_3, 8.1_0_2_2], [-8.8_6_6_4, -7.8_8_2_6, 5.5_3_2_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowercase , atol=1e-4 ) ) @slow def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) lowerCamelCase_ = BioGptForCausalLM.from_pretrained("microsoft/biogpt" ) model.to(lowercase ) torch.manual_seed(0 ) lowerCamelCase_ = tokenizer("COVID-19 is" , return_tensors="pt" ).to(lowercase ) lowerCamelCase_ = model.generate( **lowercase , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=lowercase , ) lowerCamelCase_ = tokenizer.decode(output_ids[0] , skip_special_tokens=lowercase ) lowerCamelCase_ = ( "COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the" " causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and" " territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK)," " and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and" " more than 800,000 deaths." ) self.assertEqual(lowercase , lowercase )
47
0
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase : Tuple =logging.get_logger(__name__) lowerCAmelCase : List[Any] ='''▁''' lowerCAmelCase : Dict ={'''vocab_file''': '''sentencepiece.bpe.model'''} lowerCAmelCase : int ={ '''vocab_file''': { '''facebook/xglm-564M''': '''https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model''', } } lowerCAmelCase : Tuple ={ '''facebook/xglm-564M''': 2_048, } class a_ ( _lowerCAmelCase ): __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ["input_ids", "attention_mask"] def __init__( self : Dict , lowercase : Tuple , lowercase : str="<s>" , lowercase : Optional[int]="</s>" , lowercase : Union[str, Any]="</s>" , lowercase : Any="<s>" , lowercase : Optional[Any]="<unk>" , lowercase : Union[str, Any]="<pad>" , lowercase : Optional[Dict[str, Any]] = None , **lowercase : Optional[Any] , ): """simple docstring""" lowercase_ :str = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer lowercase_ :Dict = 7 lowercase_ :Tuple = [F'<madeupword{i}>' for i in range(self.num_madeup_words )] lowercase_ :Optional[Any] = kwargs.get("additional_special_tokens" , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=lowercase , eos_token=lowercase , unk_token=lowercase , sep_token=lowercase , cls_token=lowercase , pad_token=lowercase , sp_model_kwargs=self.sp_model_kwargs , **lowercase , ) lowercase_ :Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowercase ) ) lowercase_ :List[Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab lowercase_ :Union[str, Any] = 1 # Mimic fairseq token-to-id alignment for the first 4 token lowercase_ :Union[str, Any] = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} lowercase_ :List[str] = len(self.sp_model ) lowercase_ :int = {F'<madeupword{i}>': sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(lowercase ) lowercase_ :str = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : List[str] ): """simple docstring""" lowercase_ :Any = self.__dict__.copy() lowercase_ :str = None lowercase_ :Any = self.sp_model.serialized_model_proto() return state def __setstate__( self : List[str] , lowercase : Tuple ): """simple docstring""" lowercase_ :Tuple = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowercase_ :Optional[int] = {} lowercase_ :Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def lowercase__ ( self : Tuple , lowercase : List[int] , lowercase : Optional[List[int]] = None ): """simple docstring""" if token_ids_a is None: return [self.sep_token_id] + token_ids_a lowercase_ :Union[str, Any] = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def lowercase__ ( self : int , lowercase : List[int] , lowercase : Optional[List[int]] = None , lowercase : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase , token_ids_a=lowercase , already_has_special_tokens=lowercase ) if token_ids_a is None: return [1] + ([0] * len(lowercase )) return [1] + ([0] * len(lowercase )) + [1, 1] + ([0] * len(lowercase )) def lowercase__ ( self : str , lowercase : List[int] , lowercase : Optional[List[int]] = None ): """simple docstring""" lowercase_ :Tuple = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def lowercase__ ( self : str ): """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def lowercase__ ( self : Optional[int] ): """simple docstring""" lowercase_ :Union[str, Any] = {self.convert_ids_to_tokens(lowercase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase__ ( self : str , lowercase : str ): """simple docstring""" return self.sp_model.encode(lowercase , out_type=lowercase ) def lowercase__ ( self : Dict , lowercase : str ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] lowercase_ :Optional[int] = self.sp_model.PieceToId(lowercase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowercase__ ( self : int , lowercase : str ): """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowercase__ ( self : Optional[int] , lowercase : List[Any] ): """simple docstring""" lowercase_ :Tuple = "".join(lowercase ).replace(lowercase , " " ).strip() return out_string def lowercase__ ( self : List[Any] , lowercase : str , lowercase : Optional[str] = None ): """simple docstring""" if not os.path.isdir(lowercase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return lowercase_ :Tuple = os.path.join( lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowercase ) elif not os.path.isfile(self.vocab_file ): with open(lowercase , "wb" ) as fi: lowercase_ :str = self.sp_model.serialized_model_proto() fi.write(lowercase ) return (out_vocab_file,)
223
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase : Optional[Any] ={ '''configuration_nllb_moe''': [ '''NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''NllbMoeConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase : Tuple =[ '''NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''NllbMoeForConditionalGeneration''', '''NllbMoeModel''', '''NllbMoePreTrainedModel''', '''NllbMoeTop2Router''', '''NllbMoeSparseMLP''', ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys lowerCAmelCase : Tuple =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
223
1
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging lowerCAmelCase : Optional[int] = logging.get_logger(__name__) def A_( A : Union[tf.Tensor, np.ndarray]): if isinstance(_UpperCAmelCase , np.ndarray): return list(tensor.shape) UpperCamelCase = tf.shape(_UpperCAmelCase) if tensor.shape == tf.TensorShape(_UpperCAmelCase): return dynamic UpperCamelCase = tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(_UpperCAmelCase)] def A_( A : tf.Tensor , A : Optional[int] = None , A : Optional[str] = None): return tf.nn.softmax(logits=logits + 1E-9 , axis=_UpperCAmelCase , name=_UpperCAmelCase) def A_( A : Optional[int] , A : List[Any] , A : str , A : str=1E-5 , A : Union[str, Any]=-1): # This is a very simplified functional layernorm, designed to duplicate # the functionality of PyTorch nn.functional.layer_norm when this is needed to port # models in Transformers. if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(_UpperCAmelCase , _UpperCAmelCase): raise NotImplementedError('Only 1D weight and bias tensors are supported for now, with only a single axis.') # Get mean and variance on the axis to be normalized UpperCamelCase = tf.nn.moments(_UpperCAmelCase , axes=[axis] , keepdims=_UpperCAmelCase) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis UpperCamelCase = [1] * inputs.shape.rank UpperCamelCase = shape_list(_UpperCAmelCase)[axis] UpperCamelCase = tf.reshape(_UpperCAmelCase , _UpperCAmelCase) UpperCamelCase = tf.reshape(_UpperCAmelCase , _UpperCAmelCase) # Compute layer normalization using the batch_normalization # function. UpperCamelCase = tf.nn.batch_normalization( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , offset=_UpperCAmelCase , scale=_UpperCAmelCase , variance_epsilon=_UpperCAmelCase , ) return outputs def A_( A : str , A : Dict=0 , A : int=-1): # Replicates the behavior of torch.flatten in TF # If end_dim or start_dim is negative, count them from the end if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input UpperCamelCase = tf.shape(_UpperCAmelCase) UpperCamelCase = tf.math.reduce_prod(in_shape[start_dim : end_dim + 1]) UpperCamelCase = tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0) return tf.reshape(_UpperCAmelCase , _UpperCAmelCase) def A_( A : tf.Tensor): if not isinstance(_UpperCAmelCase , tf.Tensor): UpperCamelCase = tf.convert_to_tensor(_UpperCAmelCase) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: UpperCamelCase = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: UpperCamelCase = encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) UpperCamelCase = ( tf.cast(1 , encoder_attention_mask.dtype) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def A_( A : tf.Tensor , A : int , A : str = "input_ids"): tf.debugging.assert_less( _UpperCAmelCase , tf.cast(_UpperCAmelCase , dtype=tensor.dtype) , message=( f'''The maximum value of {tensor_name} ({tf.math.reduce_max(_UpperCAmelCase)}) must be smaller than the embedding ''' f'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ) , ) def A_( A : Tuple , A : int , A : Tuple): UpperCamelCase = 6_4512 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. UpperCamelCase = [x for x in data if len(_UpperCAmelCase) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( 'The following attributes cannot be saved to HDF5 file because ' f'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' f'''bytes: {bad_attributes}''') UpperCamelCase = np.asarray(_UpperCAmelCase) UpperCamelCase = 1 UpperCamelCase = np.array_split(_UpperCAmelCase , _UpperCAmelCase) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data): num_chunks += 1 UpperCamelCase = np.array_split(_UpperCAmelCase , _UpperCAmelCase) if num_chunks > 1: for chunk_id, chunk_data in enumerate(_UpperCAmelCase): UpperCamelCase = chunk_data else: UpperCamelCase = data def A_( A : int , A : Union[str, Any]): if name in group.attrs: UpperCamelCase = [n.decode('utf8') if hasattr(_UpperCAmelCase , 'decode') else n for n in group.attrs[name]] else: UpperCamelCase = [] UpperCamelCase = 0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode('utf8') if hasattr(_UpperCAmelCase , 'decode') else n for n in group.attrs['%s%d' % (name, chunk_id)]]) chunk_id += 1 return data def A_( A : Optional[int]): def _expand_single_ad_tensor(A : Tuple): if isinstance(_UpperCAmelCase , tf.Tensor) and t.shape.rank == 1: return tf.expand_dims(_UpperCAmelCase , axis=-1) return t return tf.nest.map_structure(_expand_single_ad_tensor , _UpperCAmelCase)
350
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) lowerCAmelCase : int = { 'facebook/timesformer': 'https://huggingface.co/facebook/timesformer/resolve/main/config.json', } class SCREAMING_SNAKE_CASE__ ( snake_case_): lowerCAmelCase_ = """timesformer""" def __init__( self , A_=224 , A_=16 , A_=3 , A_=8 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.0 , A_=0.0 , A_=0.02 , A_=1e-6 , A_=True , A_="divided_space_time" , A_=0 , **A_ , )-> Union[str, Any]: '''simple docstring''' super().__init__(**A_ ) UpperCamelCase = image_size UpperCamelCase = patch_size UpperCamelCase = num_channels UpperCamelCase = num_frames UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = qkv_bias UpperCamelCase = attention_type UpperCamelCase = drop_path_rate
251
0
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = [ ('''bert.bert''', '''visual_bert'''), ('''bert.cls''', '''cls'''), ('''bert.classifier''', '''cls'''), ('''token_type_embeddings_visual''', '''visual_token_type_embeddings'''), ('''position_embeddings_visual''', '''visual_position_embeddings'''), ('''projection''', '''visual_projection'''), ] lowerCAmelCase__ = [ '''nlvr2_coco_pre_trained.th''', '''nlvr2_fine_tuned.th''', '''nlvr2_pre_trained.th''', '''vcr_coco_pre_train.th''', '''vcr_fine_tune.th''', '''vcr_pre_train.th''', '''vqa_coco_pre_trained.th''', '''vqa_fine_tuned.th''', '''vqa_pre_trained.th''', ] def snake_case_ ( A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : int = torch.load(lowerCamelCase__, map_location='''cpu''' ) return sd def snake_case_ ( A_ : str, A_ : Union[str, Any], A_ : Union[str, Any]=rename_keys_prefix ): '''simple docstring''' _lowerCamelCase : Tuple = OrderedDict() _lowerCamelCase : int = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _lowerCamelCase : List[Any] = key for name_pair in rename_keys_prefix: _lowerCamelCase : Optional[Any] = new_key.replace(name_pair[0], name_pair[1] ) _lowerCamelCase : Dict = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _lowerCamelCase : str = new_d['cls.predictions.bias'] return new_d @torch.no_grad() def snake_case_ ( A_ : int, A_ : int ): '''simple docstring''' assert ( checkpoint_path.split('''/''' )[-1] in ACCEPTABLE_CHECKPOINTS ), F'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: _lowerCamelCase : Optional[Any] = 'pretraining' if "vcr" in checkpoint_path: _lowerCamelCase : Any = {'visual_embedding_dim': 5_12} elif "vqa_advanced" in checkpoint_path: _lowerCamelCase : List[str] = {'visual_embedding_dim': 20_48} elif "vqa" in checkpoint_path: _lowerCamelCase : Tuple = {'visual_embedding_dim': 20_48} elif "nlvr" in checkpoint_path: _lowerCamelCase : List[Any] = {'visual_embedding_dim': 10_24} else: raise NotImplementedError(F'''No implementation found for `{checkpoint_path}`.''' ) else: if "vcr" in checkpoint_path: _lowerCamelCase : Tuple = {'visual_embedding_dim': 5_12} _lowerCamelCase : Any = 'multichoice' elif "vqa_advanced" in checkpoint_path: _lowerCamelCase : Tuple = {'visual_embedding_dim': 20_48} _lowerCamelCase : List[Any] = 'vqa_advanced' elif "vqa" in checkpoint_path: _lowerCamelCase : List[Any] = {'visual_embedding_dim': 20_48, 'num_labels': 31_29} _lowerCamelCase : Optional[int] = 'vqa' elif "nlvr" in checkpoint_path: _lowerCamelCase : Optional[Any] = { 'visual_embedding_dim': 10_24, 'num_labels': 2, } _lowerCamelCase : Any = 'nlvr' _lowerCamelCase : Any = VisualBertConfig(**lowerCamelCase__ ) # Load State Dict _lowerCamelCase : Dict = load_state_dict(lowerCamelCase__ ) _lowerCamelCase : Optional[Any] = get_new_dict(lowerCamelCase__, lowerCamelCase__ ) if model_type == "pretraining": _lowerCamelCase : Union[str, Any] = VisualBertForPreTraining(lowerCamelCase__ ) elif model_type == "vqa": _lowerCamelCase : str = VisualBertForQuestionAnswering(lowerCamelCase__ ) elif model_type == "nlvr": _lowerCamelCase : Dict = VisualBertForVisualReasoning(lowerCamelCase__ ) elif model_type == "multichoice": _lowerCamelCase : Union[str, Any] = VisualBertForMultipleChoice(lowerCamelCase__ ) model.load_state_dict(lowerCamelCase__ ) # Save Checkpoints Path(lowerCamelCase__ ).mkdir(exist_ok=lowerCamelCase__ ) model.save_pretrained(lowerCamelCase__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument('''orig_checkpoint_path''', type=str, help='''A path to .th on local filesystem.''') parser.add_argument('''pytorch_dump_folder_path''', type=str, help='''Path to the output PyTorch model.''') lowerCAmelCase__ = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
72
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class A_ ( SCREAMING_SNAKE_CASE ): _UpperCAmelCase : Any = ['''image_processor''', '''tokenizer'''] _UpperCAmelCase : List[Any] = '''AutoImageProcessor''' _UpperCAmelCase : Dict = '''AutoTokenizer''' def __init__( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : Optional[int]=None ,SCREAMING_SNAKE_CASE__ : List[Any]=None ,**SCREAMING_SNAKE_CASE__ : Union[str, Any]): __lowerCamelCase : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' ,SCREAMING_SNAKE_CASE__ ,) __lowerCamelCase : Union[str, Any] = kwargs.pop('feature_extractor') __lowerCamelCase : Dict = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.') if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.') super().__init__(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Dict = self.image_processor __lowerCamelCase : Optional[int] = False def __call__( self : int ,*SCREAMING_SNAKE_CASE__ : Union[str, Any] ,**SCREAMING_SNAKE_CASE__ : Union[str, Any]): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__) __lowerCamelCase : Optional[int] = kwargs.pop('images' ,SCREAMING_SNAKE_CASE__) __lowerCamelCase : Union[str, Any] = kwargs.pop('text' ,SCREAMING_SNAKE_CASE__) if len(SCREAMING_SNAKE_CASE__) > 0: __lowerCamelCase : int = args[0] __lowerCamelCase : List[str] = args[1:] if images is None and text is None: raise ValueError('You need to specify either an `images` or `text` input to process.') if images is not None: __lowerCamelCase : Optional[int] = self.image_processor(SCREAMING_SNAKE_CASE__ ,*SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__) if text is not None: __lowerCamelCase : List[Any] = self.tokenizer(SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__) if text is None: return inputs elif images is None: return encodings else: __lowerCamelCase : Optional[Any] = encodings['input_ids'] return inputs def lowerCAmelCase ( self : int ,*SCREAMING_SNAKE_CASE__ : Union[str, Any] ,**SCREAMING_SNAKE_CASE__ : Dict): return self.tokenizer.batch_decode(*SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__) def lowerCAmelCase ( self : Optional[Any] ,*SCREAMING_SNAKE_CASE__ : List[Any] ,**SCREAMING_SNAKE_CASE__ : Any): return self.tokenizer.decode(*SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__) @contextmanager def lowerCAmelCase ( self : Tuple): warnings.warn( '`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ' 'labels by using the argument `text` of the regular `__call__` method (either in the same call as ' 'your images inputs, or in a separate call.') __lowerCamelCase : List[Any] = True __lowerCamelCase : str = self.tokenizer yield __lowerCamelCase : Tuple = self.image_processor __lowerCamelCase : Tuple = False def lowerCAmelCase ( self : Any ,SCREAMING_SNAKE_CASE__ : str ,SCREAMING_SNAKE_CASE__ : int=False ,SCREAMING_SNAKE_CASE__ : List[Any]=None): if added_vocab is None: __lowerCamelCase : str = self.tokenizer.get_added_vocab() __lowerCamelCase : Union[str, Any] = {} while tokens: __lowerCamelCase : Tuple = re.search(R'<s_(.*?)>' ,SCREAMING_SNAKE_CASE__ ,re.IGNORECASE) if start_token is None: break __lowerCamelCase : Dict = start_token.group(1) __lowerCamelCase : List[str] = re.search(RF"</s_{key}>" ,SCREAMING_SNAKE_CASE__ ,re.IGNORECASE) __lowerCamelCase : Optional[int] = start_token.group() if end_token is None: __lowerCamelCase : List[Any] = tokens.replace(SCREAMING_SNAKE_CASE__ ,'') else: __lowerCamelCase : Tuple = end_token.group() __lowerCamelCase : int = re.escape(SCREAMING_SNAKE_CASE__) __lowerCamelCase : str = re.escape(SCREAMING_SNAKE_CASE__) __lowerCamelCase : Union[str, Any] = re.search(F"{start_token_escaped}(.*?){end_token_escaped}" ,SCREAMING_SNAKE_CASE__ ,re.IGNORECASE) if content is not None: __lowerCamelCase : List[Any] = content.group(1).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node __lowerCamelCase : str = self.tokenajson(SCREAMING_SNAKE_CASE__ ,is_inner_value=SCREAMING_SNAKE_CASE__ ,added_vocab=SCREAMING_SNAKE_CASE__) if value: if len(SCREAMING_SNAKE_CASE__) == 1: __lowerCamelCase : Tuple = value[0] __lowerCamelCase : int = value else: # leaf nodes __lowerCamelCase : Tuple = [] for leaf in content.split(R'<sep/>'): __lowerCamelCase : List[Any] = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": __lowerCamelCase : str = leaf[1:-2] # for categorical special tokens output[key].append(SCREAMING_SNAKE_CASE__) if len(output[key]) == 1: __lowerCamelCase : Dict = output[key][0] __lowerCamelCase : Dict = tokens[tokens.find(SCREAMING_SNAKE_CASE__) + len(SCREAMING_SNAKE_CASE__) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] ,is_inner_value=SCREAMING_SNAKE_CASE__ ,added_vocab=SCREAMING_SNAKE_CASE__) if len(SCREAMING_SNAKE_CASE__): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def lowerCAmelCase ( self : List[str]): warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' ,SCREAMING_SNAKE_CASE__ ,) return self.image_processor_class @property def lowerCAmelCase ( self : List[Any]): warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' ,SCREAMING_SNAKE_CASE__ ,) return self.image_processor
73
0
import importlib import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Union import torch from ..utils import BaseOutput lowercase_ = """scheduler_config.json""" class A_ ( __UpperCamelCase ): '''simple docstring''' __snake_case = 1 __snake_case = 2 __snake_case = 3 __snake_case = 4 __snake_case = 5 __snake_case = 6 __snake_case = 7 __snake_case = 8 __snake_case = 9 __snake_case = 10 __snake_case = 11 __snake_case = 12 __snake_case = 13 __snake_case = 14 @dataclass class A_ ( __UpperCamelCase ): '''simple docstring''' __snake_case = 42 class A_ : '''simple docstring''' __snake_case = SCHEDULER_CONFIG_NAME __snake_case = [] __snake_case = True @classmethod def _snake_case ( cls: int , a: Optional[int] = None , a: List[Any] = None , a: Union[str, Any]=False , **a: Any , ) -> Dict: __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : str = cls.load_config( pretrained_model_name_or_path=__lowerCAmelCase , subfolder=__lowerCAmelCase , return_unused_kwargs=__lowerCAmelCase , return_commit_hash=__lowerCAmelCase , **__lowerCAmelCase , ) return cls.from_config(__lowerCAmelCase , return_unused_kwargs=__lowerCAmelCase , **__lowerCAmelCase ) def _snake_case ( self: Optional[Any] , a: List[str] , a: Optional[int] = False , **a: str ) -> Tuple: self.save_config(save_directory=__lowerCAmelCase , push_to_hub=__lowerCAmelCase , **__lowerCAmelCase ) @property def _snake_case ( self: Optional[Any] ) -> List[str]: return self._get_compatibles() @classmethod def _snake_case ( cls: Optional[Any] ) -> Dict: __lowerCamelCase : Optional[int] = list(set([cls.__name__] + cls._compatibles ) ) __lowerCamelCase : Optional[Any] = importlib.import_module(__name__.split('.' )[0] ) __lowerCamelCase : str = [ getattr(__lowerCAmelCase , __lowerCAmelCase ) for c in compatible_classes_str if hasattr(__lowerCAmelCase , __lowerCAmelCase ) ] return compatible_classes
360
import warnings from .generation import TFGenerationMixin class A_ ( __UpperCamelCase ): '''simple docstring''' warnings.warn( """Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will """ """be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead.""" , __UpperCamelCase , )
194
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available snake_case_ : Optional[Any] = { 'configuration_longt5': ['LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LongT5Config', 'LongT5OnnxConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : str = [ 'LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST', 'LongT5EncoderModel', 'LongT5ForConditionalGeneration', 'LongT5Model', 'LongT5PreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Dict = [ 'FlaxLongT5ForConditionalGeneration', 'FlaxLongT5Model', 'FlaxLongT5PreTrainedModel', ] if TYPE_CHECKING: from .configuration_longta import LONGT5_PRETRAINED_CONFIG_ARCHIVE_MAP, LongTaConfig, LongTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longta import ( LONGT5_PRETRAINED_MODEL_ARCHIVE_LIST, LongTaEncoderModel, LongTaForConditionalGeneration, LongTaModel, LongTaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_longta import ( FlaxLongTaForConditionalGeneration, FlaxLongTaModel, FlaxLongTaPreTrainedModel, ) else: import sys snake_case_ : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
83
'''simple docstring''' def UpperCamelCase_ ( snake_case_ : list[int] , snake_case_ : list[int] ) -> tuple[float, float]: '''simple docstring''' if not len(snake_case_ ) == len(snake_case_ ) == 3: raise ValueError("""Please enter a valid equation.""" ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError("""Both a & b of two equations can't be zero.""" ) # Extract the coefficients __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = equationa __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = equationa # Calculate the determinants of the matrices __lowerCAmelCase = aa * ba - aa * ba __lowerCAmelCase = ca * ba - ca * ba __lowerCAmelCase = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError("""Infinite solutions. (Consistent system)""" ) else: raise ValueError("""No solution. (Inconsistent system)""" ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: __lowerCAmelCase = determinant_x / determinant __lowerCAmelCase = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
229
0
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _UpperCAmelCase : List[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name class __lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase): @register_to_config def __init__( self: Any , _lowerCAmelCase: bool , _lowerCAmelCase: Optional[int] = None , _lowerCAmelCase: Optional[int] = None ): super().__init__() lowercase :Union[str, Any] = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" lowercase :Optional[int] = torch.zeros(_lowerCAmelCase , _lowerCAmelCase ) else: lowercase :Any = None lowercase :List[Any] = torch.nn.Parameter(_lowerCAmelCase ) class __lowerCAmelCase ( lowerCAmelCase): _a = 42 _a = 42 _a = 42 _a = 42 _a = 42 _a = 42 def __init__( self: Any , _lowerCAmelCase: VQModel , _lowerCAmelCase: CLIPTextModel , _lowerCAmelCase: CLIPTokenizer , _lowerCAmelCase: TransformeraDModel , _lowerCAmelCase: VQDiffusionScheduler , _lowerCAmelCase: LearnedClassifierFreeSamplingEmbeddings , ): super().__init__() self.register_modules( vqvae=_lowerCAmelCase , transformer=_lowerCAmelCase , text_encoder=_lowerCAmelCase , tokenizer=_lowerCAmelCase , scheduler=_lowerCAmelCase , learned_classifier_free_sampling_embeddings=_lowerCAmelCase , ) def SCREAMING_SNAKE_CASE ( self: str , _lowerCAmelCase: Tuple , _lowerCAmelCase: List[Any] , _lowerCAmelCase: Dict ): lowercase :Union[str, Any] = len(_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else 1 # get prompt text embeddings lowercase :Tuple = self.tokenizer( _lowerCAmelCase , padding="max_length" , max_length=self.tokenizer.model_max_length , return_tensors="pt" , ) lowercase :Any = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: lowercase :Dict = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( "The following part of your input was truncated because CLIP can only handle sequences up to" F" {self.tokenizer.model_max_length} tokens: {removed_text}" ) lowercase :str = text_input_ids[:, : self.tokenizer.model_max_length] lowercase :Union[str, Any] = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 lowercase :List[Any] = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=_lowerCAmelCase ) # duplicate text embeddings for each generation per prompt lowercase :List[Any] = prompt_embeds.repeat_interleave(_lowerCAmelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: lowercase :Union[str, Any] = self.learned_classifier_free_sampling_embeddings.embeddings lowercase :int = negative_prompt_embeds.unsqueeze(0 ).repeat(_lowerCAmelCase , 1 , 1 ) else: lowercase :Optional[int] = [""] * batch_size lowercase :Any = text_input_ids.shape[-1] lowercase :int = self.tokenizer( _lowerCAmelCase , padding="max_length" , max_length=_lowerCAmelCase , truncation=_lowerCAmelCase , return_tensors="pt" , ) lowercase :List[str] = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings lowercase :Optional[Any] = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=_lowerCAmelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method lowercase :Any = negative_prompt_embeds.shape[1] lowercase :str = negative_prompt_embeds.repeat(1 , _lowerCAmelCase , 1 ) lowercase :Union[str, Any] = negative_prompt_embeds.view(batch_size * num_images_per_prompt , _lowerCAmelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowercase :Any = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self: Tuple , _lowerCAmelCase: Union[str, List[str]] , _lowerCAmelCase: int = 1_00 , _lowerCAmelCase: float = 5.0 , _lowerCAmelCase: float = 1.0 , _lowerCAmelCase: int = 1 , _lowerCAmelCase: Optional[Union[torch.Generator, List[torch.Generator]]] = None , _lowerCAmelCase: Optional[torch.FloatTensor] = None , _lowerCAmelCase: Optional[str] = "pil" , _lowerCAmelCase: bool = True , _lowerCAmelCase: Optional[Callable[[int, int, torch.FloatTensor], None]] = None , _lowerCAmelCase: int = 1 , ): if isinstance(_lowerCAmelCase , _lowerCAmelCase ): lowercase :Any = 1 elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): lowercase :Any = len(_lowerCAmelCase ) else: raise ValueError(F"`prompt` has to be of type `str` or `list` but is {type(_lowerCAmelCase )}" ) lowercase :List[str] = batch_size * num_images_per_prompt lowercase :Optional[Any] = guidance_scale > 1.0 lowercase :Union[str, Any] = self._encode_prompt(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(_lowerCAmelCase , _lowerCAmelCase ) or callback_steps <= 0) ): raise ValueError( F"`callback_steps` has to be a positive integer but is {callback_steps} of type" F" {type(_lowerCAmelCase )}." ) # get the initial completely masked latents unless the user supplied it lowercase :Dict = (batch_size, self.transformer.num_latent_pixels) if latents is None: lowercase :Union[str, Any] = self.transformer.num_vector_embeds - 1 lowercase :Union[str, Any] = torch.full(_lowerCAmelCase , _lowerCAmelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(F"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( "Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0," F" {self.transformer.num_vector_embeds - 1} (inclusive)." ) lowercase :Any = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(_lowerCAmelCase , device=self.device ) lowercase :Optional[int] = self.scheduler.timesteps.to(self.device ) lowercase :Union[str, Any] = latents for i, t in enumerate(self.progress_bar(_lowerCAmelCase ) ): # expand the sample if we are doing classifier free guidance lowercase :Union[str, Any] = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` lowercase :Union[str, Any] = self.transformer(_lowerCAmelCase , encoder_hidden_states=_lowerCAmelCase , timestep=_lowerCAmelCase ).sample if do_classifier_free_guidance: lowercase , lowercase :Optional[int] = model_output.chunk(2 ) lowercase :Tuple = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(_lowerCAmelCase , dim=1 , keepdim=_lowerCAmelCase ) lowercase :Tuple = self.truncate(_lowerCAmelCase , _lowerCAmelCase ) # remove `log(0)`'s (`-inf`s) lowercase :Tuple = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 lowercase :int = self.scheduler.step(_lowerCAmelCase , timestep=_lowerCAmelCase , sample=_lowerCAmelCase , generator=_lowerCAmelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase :List[str] = self.vqvae.config.vq_embed_dim lowercase :str = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) lowercase :Tuple = self.vqvae.quantize.get_codebook_entry(_lowerCAmelCase , shape=_lowerCAmelCase ) lowercase :Optional[Any] = self.vqvae.decode(_lowerCAmelCase , force_not_quantize=_lowerCAmelCase ).sample lowercase :int = (image / 2 + 0.5).clamp(0 , 1 ) lowercase :Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowercase :Optional[Any] = self.numpy_to_pil(_lowerCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: torch.FloatTensor , _lowerCAmelCase: float ): lowercase , lowercase :Optional[int] = torch.sort(_lowerCAmelCase , 1 , descending=_lowerCAmelCase ) lowercase :List[Any] = torch.exp(_lowerCAmelCase ) lowercase :List[Any] = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out lowercase :Optional[Any] = torch.full_like(keep_mask[:, 0:1, :] , _lowerCAmelCase ) lowercase :Union[str, Any] = torch.cat((all_true, keep_mask) , dim=1 ) lowercase :Optional[int] = keep_mask[:, :-1, :] lowercase :Any = keep_mask.gather(1 , indices.argsort(1 ) ) lowercase :List[str] = log_p_x_0.clone() lowercase :Optional[Any] = -torch.inf # -inf = log(0) return rv
158
import logging import os import threading import time try: import warnings except ImportError: _UpperCAmelCase : List[str] = None try: import msvcrt except ImportError: _UpperCAmelCase : Tuple = None try: import fcntl except ImportError: _UpperCAmelCase : Optional[Any] = None # Backward compatibility # ------------------------------------------------ try: TimeoutError except NameError: _UpperCAmelCase : Tuple = OSError # Data # ------------------------------------------------ _UpperCAmelCase : Optional[int] = [ "Timeout", "BaseFileLock", "WindowsFileLock", "UnixFileLock", "SoftFileLock", "FileLock", ] _UpperCAmelCase : Optional[Any] = "3.0.12" _UpperCAmelCase : int = None def UpperCAmelCase__ ( ): global _logger lowercase :List[str] = _logger or logging.getLogger(__name__ ) return _logger class __lowerCAmelCase ( lowerCAmelCase): def __init__( self: int , _lowerCAmelCase: Dict ): lowercase :Any = lock_file return None def __str__( self: Dict ): lowercase :str = F"The file lock '{self.lock_file}' could not be acquired." return temp class __lowerCAmelCase : def __init__( self: Tuple , _lowerCAmelCase: Any ): lowercase :Optional[Any] = lock return None def __enter__( self: List[Any] ): return self.lock def __exit__( self: Dict , _lowerCAmelCase: Optional[Any] , _lowerCAmelCase: List[Any] , _lowerCAmelCase: Optional[int] ): self.lock.release() return None class __lowerCAmelCase : def __init__( self: Optional[Any] , _lowerCAmelCase: Union[str, Any] , _lowerCAmelCase: Tuple=-1 , _lowerCAmelCase: int=None ): lowercase :Any = max_filename_length if max_filename_length is not None else 2_55 # Hash the filename if it's too long lowercase :int = self.hash_filename_if_too_long(_lowerCAmelCase , _lowerCAmelCase ) # The path to the lock file. lowercase :List[Any] = lock_file # The file descriptor for the *_lock_file* as it is returned by the # os.open() function. # This file lock is only NOT None, if the object currently holds the # lock. lowercase :Any = None # The default timeout value. lowercase :Any = timeout # We use this lock primarily for the lock counter. lowercase :Optional[int] = threading.Lock() # The lock counter is used for implementing the nested locking # mechanism. Whenever the lock is acquired, the counter is increased and # the lock is only released, when this value is 0 again. lowercase :Optional[int] = 0 return None @property def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): return self._lock_file @property def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): return self._timeout @timeout.setter def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: List[str] ): lowercase :Tuple = float(_lowerCAmelCase ) return None def SCREAMING_SNAKE_CASE ( self: int ): raise NotImplementedError() def SCREAMING_SNAKE_CASE ( self: int ): raise NotImplementedError() @property def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): return self._lock_file_fd is not None def SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _lowerCAmelCase: List[Any]=None , _lowerCAmelCase: Union[str, Any]=0.05 ): # Use the default timeout, if no timeout is provided. if timeout is None: lowercase :List[str] = self.timeout # Increment the number right at the beginning. # We can still undo it, if something fails. with self._thread_lock: self._lock_counter += 1 lowercase :Any = id(self ) lowercase :Optional[int] = self._lock_file lowercase :Optional[Any] = time.time() try: while True: with self._thread_lock: if not self.is_locked: logger().debug(F"Attempting to acquire lock {lock_id} on {lock_filename}" ) self._acquire() if self.is_locked: logger().debug(F"Lock {lock_id} acquired on {lock_filename}" ) break elif timeout >= 0 and time.time() - start_time > timeout: logger().debug(F"Timeout on acquiring lock {lock_id} on {lock_filename}" ) raise Timeout(self._lock_file ) else: logger().debug( F"Lock {lock_id} not acquired on {lock_filename}, waiting {poll_intervall} seconds ..." ) time.sleep(_lowerCAmelCase ) except: # noqa # Something did go wrong, so decrement the counter. with self._thread_lock: lowercase :Union[str, Any] = max(0 , self._lock_counter - 1 ) raise return _Acquire_ReturnProxy(lock=self ) def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: Tuple=False ): with self._thread_lock: if self.is_locked: self._lock_counter -= 1 if self._lock_counter == 0 or force: lowercase :Union[str, Any] = id(self ) lowercase :str = self._lock_file logger().debug(F"Attempting to release lock {lock_id} on {lock_filename}" ) self._release() lowercase :List[str] = 0 logger().debug(F"Lock {lock_id} released on {lock_filename}" ) return None def __enter__( self: Tuple ): self.acquire() return self def __exit__( self: Union[str, Any] , _lowerCAmelCase: List[Any] , _lowerCAmelCase: List[str] , _lowerCAmelCase: Dict ): self.release() return None def __del__( self: Optional[Any] ): self.release(force=_lowerCAmelCase ) return None def SCREAMING_SNAKE_CASE ( self: Union[str, Any] , _lowerCAmelCase: str , _lowerCAmelCase: int ): lowercase :Union[str, Any] = os.path.basename(_lowerCAmelCase ) if len(_lowerCAmelCase ) > max_length and max_length > 0: lowercase :Dict = os.path.dirname(_lowerCAmelCase ) lowercase :Any = str(hash(_lowerCAmelCase ) ) lowercase :Union[str, Any] = filename[: max_length - len(_lowerCAmelCase ) - 8] + "..." + hashed_filename + ".lock" return os.path.join(_lowerCAmelCase , _lowerCAmelCase ) else: return path class __lowerCAmelCase ( lowerCAmelCase): def __init__( self: int , _lowerCAmelCase: int , _lowerCAmelCase: Optional[Any]=-1 , _lowerCAmelCase: List[Any]=None ): from .file_utils import relative_to_absolute_path super().__init__(_lowerCAmelCase , timeout=_lowerCAmelCase , max_filename_length=_lowerCAmelCase ) lowercase :Optional[int] = "\\\\?\\" + relative_to_absolute_path(self.lock_file ) def SCREAMING_SNAKE_CASE ( self: Any ): lowercase :int = os.O_RDWR | os.O_CREAT | os.O_TRUNC try: lowercase :Tuple = os.open(self._lock_file , _lowerCAmelCase ) except OSError: pass else: try: msvcrt.locking(_lowerCAmelCase , msvcrt.LK_NBLCK , 1 ) except OSError: os.close(_lowerCAmelCase ) else: lowercase :Any = fd return None def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): lowercase :Any = self._lock_file_fd lowercase :Tuple = None msvcrt.locking(_lowerCAmelCase , msvcrt.LK_UNLCK , 1 ) os.close(_lowerCAmelCase ) try: os.remove(self._lock_file ) # Probably another instance of the application # that acquired the file lock. except OSError: pass return None class __lowerCAmelCase ( lowerCAmelCase): def __init__( self: str , _lowerCAmelCase: Tuple , _lowerCAmelCase: Dict=-1 , _lowerCAmelCase: Tuple=None ): lowercase :List[str] = os.statvfs(os.path.dirname(_lowerCAmelCase ) ).f_namemax super().__init__(_lowerCAmelCase , timeout=_lowerCAmelCase , max_filename_length=_lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self: int ): lowercase :Any = os.O_RDWR | os.O_CREAT | os.O_TRUNC lowercase :Optional[int] = os.open(self._lock_file , _lowerCAmelCase ) try: fcntl.flock(_lowerCAmelCase , fcntl.LOCK_EX | fcntl.LOCK_NB ) except OSError: os.close(_lowerCAmelCase ) else: lowercase :Optional[Any] = fd return None def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): # Do not remove the lockfile: # # https://github.com/benediktschmitt/py-filelock/issues/31 # https://stackoverflow.com/questions/17708885/flock-removing-locked-file-without-race-condition lowercase :Dict = self._lock_file_fd lowercase :Union[str, Any] = None fcntl.flock(_lowerCAmelCase , fcntl.LOCK_UN ) os.close(_lowerCAmelCase ) return None class __lowerCAmelCase ( lowerCAmelCase): def SCREAMING_SNAKE_CASE ( self: List[Any] ): lowercase :str = os.O_WRONLY | os.O_CREAT | os.O_EXCL | os.O_TRUNC try: lowercase :List[Any] = os.open(self._lock_file , _lowerCAmelCase ) except OSError: pass else: lowercase :int = fd return None def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): os.close(self._lock_file_fd ) lowercase :int = None try: os.remove(self._lock_file ) # The file is already deleted and that's what we want. except OSError: pass return None _UpperCAmelCase : Tuple = None if msvcrt: _UpperCAmelCase : str = WindowsFileLock elif fcntl: _UpperCAmelCase : List[Any] = UnixFileLock else: _UpperCAmelCase : Optional[int] = SoftFileLock if warnings is not None: warnings.warn("only soft file lock is available")
158
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class A__ ( metaclass=_lowerCamelCase): A_ : List[str] = ['flax', 'transformers'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] ) class A__ ( metaclass=_lowerCamelCase): A_ : List[Any] = ['flax', 'transformers'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] ) class A__ ( metaclass=_lowerCamelCase): A_ : str = ['flax', 'transformers'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] ) class A__ ( metaclass=_lowerCamelCase): A_ : Any = ['flax', 'transformers'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax', 'transformers'] )
86
"""simple docstring""" from __future__ import annotations from scipy.special import comb # type: ignore class _A : def __init__( self , __lowerCAmelCase ): """simple docstring""" lowercase = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. lowercase = len(__lowerCAmelCase ) - 1 def A__ ( self , __lowerCAmelCase ): """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." lowercase = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __lowerCAmelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__lowerCAmelCase ) , 5 ) == 1 return output_values def A__ ( self , __lowerCAmelCase ): """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." lowercase = self.basis_function(__lowerCAmelCase ) lowercase = 0.0 lowercase = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def A__ ( self , __lowerCAmelCase = 0.0_1 ): """simple docstring""" from matplotlib import pyplot as plt # type: ignore lowercase = [] # x coordinates of points to plot lowercase = [] # y coordinates of points to plot lowercase = 0.0 while t <= 1: lowercase = self.bezier_curve_function(__lowerCAmelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size lowercase = [i[0] for i in self.list_of_points] lowercase = [i[1] for i in self.list_of_points] plt.plot( __lowerCAmelCase , __lowerCAmelCase , color="""blue""" , label="""Curve of Degree """ + str(self.degree ) , ) plt.scatter(__lowerCAmelCase , __lowerCAmelCase , color="""red""" , label="""Control Points""" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
197
0
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A : List[str] = {'configuration_mra': ['MRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MraConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : Tuple = [ 'MRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'MraForMaskedLM', 'MraForMultipleChoice', 'MraForQuestionAnswering', 'MraForSequenceClassification', 'MraForTokenClassification', 'MraLayer', 'MraModel', 'MraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mra import MRA_PRETRAINED_CONFIG_ARCHIVE_MAP, MraConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mra import ( MRA_PRETRAINED_MODEL_ARCHIVE_LIST, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraLayer, MraModel, MraPreTrainedModel, ) else: import sys __A : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure)
360
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __A : List[Any] = logging.get_logger(__name__) __A : str = { 'google/canine-s': 'https://huggingface.co/google/canine-s/resolve/main/config.json', # See all CANINE models at https://huggingface.co/models?filter=canine } class __UpperCamelCase ( lowercase__ ): lowercase : List[Any] = 'canine' def __init__( self :Optional[int] ,_UpperCamelCase :Dict=7_6_8 ,_UpperCamelCase :Union[str, Any]=1_2 ,_UpperCamelCase :int=1_2 ,_UpperCamelCase :int=3_0_7_2 ,_UpperCamelCase :int="gelu" ,_UpperCamelCase :Any=0.1 ,_UpperCamelCase :int=0.1 ,_UpperCamelCase :Any=1_6_3_8_4 ,_UpperCamelCase :Tuple=1_6 ,_UpperCamelCase :List[str]=0.02 ,_UpperCamelCase :Any=1E-1_2 ,_UpperCamelCase :Tuple=0 ,_UpperCamelCase :List[str]=0xE_0_0_0 ,_UpperCamelCase :Optional[Any]=0xE_0_0_1 ,_UpperCamelCase :str=4 ,_UpperCamelCase :Optional[int]=4 ,_UpperCamelCase :str=8 ,_UpperCamelCase :int=1_6_3_8_4 ,_UpperCamelCase :int=1_2_8 ,**_UpperCamelCase :str ,): super().__init__(pad_token_id=_UpperCamelCase ,bos_token_id=_UpperCamelCase ,eos_token_id=_UpperCamelCase ,**_UpperCamelCase ) snake_case_ : List[str] = max_position_embeddings snake_case_ : Union[str, Any] = hidden_size snake_case_ : Dict = num_hidden_layers snake_case_ : Optional[int] = num_attention_heads snake_case_ : Tuple = intermediate_size snake_case_ : str = hidden_act snake_case_ : Union[str, Any] = hidden_dropout_prob snake_case_ : Dict = attention_probs_dropout_prob snake_case_ : Optional[Any] = initializer_range snake_case_ : Optional[int] = type_vocab_size snake_case_ : List[str] = layer_norm_eps # Character config: snake_case_ : Any = downsampling_rate snake_case_ : List[str] = upsampling_kernel_size snake_case_ : int = num_hash_functions snake_case_ : Tuple = num_hash_buckets snake_case_ : Tuple = local_transformer_stride
8
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __UpperCAmelCase : Union[str, Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : str = ["BartphoTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys __UpperCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
111
__UpperCAmelCase : int = [ "Audio", "Array2D", "Array3D", "Array4D", "Array5D", "ClassLabel", "Features", "Sequence", "Value", "Image", "Translation", "TranslationVariableLanguages", ] from .audio import Audio from .features import ArrayaD, ArrayaD, ArrayaD, ArrayaD, ClassLabel, Features, Sequence, Value from .image import Image from .translation import Translation, TranslationVariableLanguages
111
1
'''simple docstring''' import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py _snake_case = 'src/diffusers' # Matches is_xxx_available() _snake_case = re.compile(r'is\_([a-z_]*)_available\(\)') # Matches from xxx import bla _snake_case = re.compile(r'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') _snake_case = '\n{0} = None\n' _snake_case = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n\n @classmethod\n def from_config(cls, *args, **kwargs):\n requires_backends(cls, {1})\n\n @classmethod\n def from_pretrained(cls, *args, **kwargs):\n requires_backends(cls, {1})\n' _snake_case = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' def _A ( snake_case ) -> int: _lowercase : Dict = _re_backend.findall(snake_case ) if len(snake_case ) == 0: return None return "_and_".join(snake_case ) def _A ( ) -> int: with open(os.path.join(snake_case , "__init__.py" ) , "r" , encoding="utf-8" , newline="\n" ) as f: _lowercase : List[str] = f.readlines() # Get to the point we do the actual imports for type checking _lowercase : Any = 0 _lowercase : List[str] = {} # Go through the end of the file while line_index < len(snake_case ): # If the line contains is_backend_available, we grab all objects associated with the `else` block _lowercase : List[Any] = find_backend(lines[line_index] ) if backend is not None: while not lines[line_index].startswith("else:" ): line_index += 1 line_index += 1 _lowercase : Tuple = [] # Until we unindent, add backend objects to the list while line_index < len(snake_case ) and len(lines[line_index] ) > 1: _lowercase : Dict = lines[line_index] _lowercase : Union[str, Any] = _re_single_line_import.search(snake_case ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(", " ) ) elif line.startswith(" " * 8 ): objects.append(line[8:-2] ) line_index += 1 if len(snake_case ) > 0: _lowercase : Dict = objects else: line_index += 1 return backend_specific_objects def _A ( snake_case , snake_case ) -> Optional[int]: if name.isupper(): return DUMMY_CONSTANT.format(snake_case ) elif name.islower(): return DUMMY_FUNCTION.format(snake_case , snake_case ) else: return DUMMY_CLASS.format(snake_case , snake_case ) def _A ( snake_case=None ) -> Any: if backend_specific_objects is None: _lowercase : Optional[int] = read_init() # For special correspondence backend to module name as used in the function requires_modulename _lowercase : List[str] = {} for backend, objects in backend_specific_objects.items(): _lowercase : int = "[" + ", ".join(F'''"{b}"''' for b in backend.split("_and_" ) ) + "]" _lowercase : Optional[int] = "# This file is autogenerated by the command `make fix-copies`, do not edit.\n" dummy_file += "from ..utils import DummyObject, requires_backends\n\n" dummy_file += "\n".join([create_dummy_object(snake_case , snake_case ) for o in objects] ) _lowercase : Optional[int] = dummy_file return dummy_files def _A ( snake_case=False ) -> Optional[int]: _lowercase : List[Any] = create_dummy_files() # For special correspondence backend to shortcut as used in utils/dummy_xxx_objects.py _lowercase : Optional[Any] = {"torch": "pt"} # Locate actual dummy modules and read their content. _lowercase : int = os.path.join(snake_case , "utils" ) _lowercase : Optional[Any] = { backend: os.path.join(snake_case , F'''dummy_{short_names.get(snake_case , snake_case )}_objects.py''' ) for backend in dummy_files.keys() } _lowercase : Optional[int] = {} for backend, file_path in dummy_file_paths.items(): if os.path.isfile(snake_case ): with open(snake_case , "r" , encoding="utf-8" , newline="\n" ) as f: _lowercase : Optional[Any] = f.read() else: _lowercase : Any = "" for backend in dummy_files.keys(): if dummy_files[backend] != actual_dummies[backend]: if overwrite: print( F'''Updating diffusers.utils.dummy_{short_names.get(snake_case , snake_case )}_objects.py as the main ''' "__init__ has new objects." ) with open(dummy_file_paths[backend] , "w" , encoding="utf-8" , newline="\n" ) as f: f.write(dummy_files[backend] ) else: raise ValueError( "The main __init__ has objects that are not present in " F'''diffusers.utils.dummy_{short_names.get(snake_case , snake_case )}_objects.py. Run `make fix-copies` ''' "to fix this." ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') _snake_case = parser.parse_args() check_dummies(args.fix_and_overwrite)
199
'''simple docstring''' import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def _A ( snake_case=32 , snake_case=10 , snake_case=1_00 , snake_case=10_26 , snake_case=True , snake_case="data/tokenized_stories_train_wikitext103.jbl" , snake_case="igf_context_pairs.jbl" , ) -> Optional[int]: set_seed(3 ) # generate train_data and objective_set _lowercase , _lowercase : List[str] = generate_datasets( snake_case , snake_case , number=snake_case , min_len=10_26 , trim=snake_case ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? _lowercase : int = torch.device("cuda:0" if torch.cuda.is_available() else "cpu" ) # load pretrained model _lowercase : str = load_gpta("gpt2" ).to(snake_case ) print("computing perplexity on objective set" ) _lowercase : Dict = compute_perplexity(snake_case , snake_case , snake_case ).item() print("perplexity on objective set:" , snake_case ) # collect igf pairs and save to file demo.jbl collect_objective_set(snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def _A ( snake_case , snake_case=15 , snake_case=1_28 , snake_case=1_00 , snake_case="igf_model.pt" , ) -> Optional[Any]: set_seed(42 ) # Load pre-trained model _lowercase : Tuple = GPTaLMHeadModel.from_pretrained("gpt2" ) # Initialize secondary learner to use embedding weights of model _lowercase : Any = SecondaryLearner(snake_case ) # Train secondary learner _lowercase : Any = train_secondary_learner( snake_case , snake_case , max_epochs=snake_case , batch_size=snake_case , eval_freq=1_00 , igf_model_path=snake_case , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def _A ( snake_case , snake_case , snake_case , snake_case=32 , snake_case=10_00 , snake_case=16 , snake_case=1.0 , snake_case=recopy_gpta , snake_case=None , snake_case=10 , snake_case="gpt2_finetuned.pt" , ) -> Dict: _lowercase : str = torch.device("cuda:0" if torch.cuda.is_available() else "cpu" ) _lowercase : int = RandomSampler(snake_case ) _lowercase : int = DataLoader(snake_case , sampler=snake_case ) _lowercase : Tuple = max_steps // (len(snake_case )) + 1 _lowercase : Dict = 0 _lowercase : Union[str, Any] = torch.zeros((1, context_len) , dtype=torch.long , device=snake_case ) _lowercase , _lowercase , _lowercase : Union[str, Any] = recopy_model(snake_case , snake_case , snake_case ) model.train() if secondary_learner is not None: secondary_learner.to(snake_case ) secondary_learner.eval() _lowercase : Optional[Any] = [] _lowercase : Tuple = 0 _lowercase : int = [] _lowercase : Optional[Any] = [] # Compute the performance of the transformer model at the beginning _lowercase : Dict = compute_perplexity(snake_case , snake_case , snake_case ) test_perps.append(snake_case ) print("Test perplexity, step" , snake_case , ":" , snake_case ) for epoch in range(int(snake_case ) ): for step, example in enumerate(snake_case ): torch.cuda.empty_cache() _lowercase : Optional[Any] = random.randint(0 , example.size(2 ) - context_len - 1 ) _lowercase : Tuple = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() _lowercase : Tuple = model(snake_case , labels=snake_case ) _lowercase : List[Any] = True if secondary_learner is not None: _lowercase : Dict = secondary_learner.forward( torch.tensor(snake_case , dtype=torch.long , device=snake_case ).unsqueeze(0 ) )[0].item() observed_qs.append(float(snake_case ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: _lowercase : Optional[Any] = -1 if predicted_q < threshold: _lowercase : List[str] = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) _lowercase : Dict = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() _lowercase : Optional[Any] = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: _lowercase : Optional[Any] = compute_perplexity(snake_case , snake_case , snake_case ) test_perps.append(snake_case ) print("Test perplexity, step" , snake_case , ":" , snake_case ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , snake_case ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def _A ( ) -> Union[str, Any]: _lowercase : Optional[Any] = argparse.ArgumentParser(description="Fine-tune a transformer model with IGF on a language modeling task" ) # Required parameters parser.add_argument( "--data_dir" , default=snake_case , type=snake_case , required=snake_case , help="The input data dir. Should contain data files for WikiText." , ) parser.add_argument( "--model_name_or_path" , default=snake_case , type=snake_case , required=snake_case , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--data_file" , type=snake_case , default=snake_case , help=( "A jbl file containing tokenized data which can be split as objective dataset, " "train_dataset and test_dataset." ) , ) parser.add_argument( "--igf_data_file" , type=snake_case , default=snake_case , help="A jbl file containing the context and information gain pairs to train secondary learner." , ) parser.add_argument( "--output_dir" , default=snake_case , type=snake_case , required=snake_case , help="The output directory where the final fine-tuned model is stored." , ) parser.add_argument( "--tokenizer_name" , default=snake_case , type=snake_case , help="Pretrained tokenizer name or path if not the same as model_name" , ) parser.add_argument("--seed" , type=snake_case , default=snake_case , help="A seed for reproducible training." ) parser.add_argument( "--context_len" , default=32 , type=snake_case , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--size_objective_set" , default=1_00 , type=snake_case , help="number of articles that are long enough to be used as our objective set" , ) parser.add_argument( "--eval_freq" , default=1_00 , type=snake_case , help="secondary model evaluation is triggered at eval_freq" ) parser.add_argument("--max_steps" , default=10_00 , type=snake_case , help="To calculate training epochs" ) parser.add_argument( "--secondary_learner_batch_size" , default=1_28 , type=snake_case , help="batch size of training data for secondary learner" , ) parser.add_argument( "--batch_size" , default=16 , type=snake_case , help="batch size of training data of language model(gpt2) " ) parser.add_argument( "--eval_interval" , default=10 , type=snake_case , help=( "decay the selectivity of our secondary learner filter from" "1 standard deviation above average to 1 below average after 10 batches" ) , ) parser.add_argument( "--number" , default=1_00 , type=snake_case , help="The number of examples split to be used as objective_set/test_data" ) parser.add_argument( "--min_len" , default=10_26 , type=snake_case , help="The minimum length of the article to be used as objective set" ) parser.add_argument( "--secondary_learner_max_epochs" , default=15 , type=snake_case , help="number of epochs to train secondary learner" ) parser.add_argument("--trim" , default=snake_case , type=snake_case , help="truncate the example if it exceeds context length" ) parser.add_argument( "--threshold" , default=1.0 , type=snake_case , help=( "The threshold value used by secondary learner to filter the train_data and allow only" " informative data as input to the model" ) , ) parser.add_argument("--finetuned_model_name" , default="gpt2_finetuned.pt" , type=snake_case , help="finetuned_model_name" ) parser.add_argument( "--recopy_model" , default=snake_case , type=snake_case , help="Reset the model to the original pretrained GPT-2 weights after each iteration" , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=1_00 , min_len=10_26 , trim=snake_case , data_file="data/tokenized_stories_train_wikitext103.jbl" , igf_data_file="igf_context_pairs.jbl" , ) # Load train data for secondary learner _lowercase : Any = joblib.load("data/IGF_values.jbl" ) # Train secondary learner _lowercase : Union[str, Any] = training_secondary_learner( snake_case , secondary_learner_max_epochs=15 , secondary_learner_batch_size=1_28 , eval_freq=1_00 , igf_model_path="igf_model.pt" , ) # load pretrained gpt2 model _lowercase : Optional[Any] = GPTaLMHeadModel.from_pretrained("gpt2" ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model _lowercase , _lowercase : Optional[Any] = generate_datasets( context_len=32 , file="data/tokenized_stories_train_wikitext103.jbl" , number=1_00 , min_len=10_26 , trim=snake_case ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( snake_case , snake_case , snake_case , context_len=32 , max_steps=10_00 , batch_size=16 , threshold=1.0 , recopy_model=snake_case , secondary_learner=snake_case , eval_interval=10 , finetuned_model_name="gpt2_finetuned.pt" , ) if __name__ == "__main__": main()
199
1
'''simple docstring''' import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _lowerCAmelCase ( __UpperCAmelCase ): __SCREAMING_SNAKE_CASE : List[Any] = ['image_processor', 'tokenizer'] __SCREAMING_SNAKE_CASE : Optional[Any] = 'OwlViTImageProcessor' __SCREAMING_SNAKE_CASE : List[str] = ('CLIPTokenizer', 'CLIPTokenizerFast') def __init__(self , lowercase=None , lowercase=None , **lowercase ): A_ : List[str] = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , lowercase , ) A_ : List[Any] = kwargs.pop("""feature_extractor""" ) A_ : Any = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(lowercase , lowercase ) def __call__(self , lowercase=None , lowercase=None , lowercase=None , lowercase="max_length" , lowercase="np" , **lowercase ): if text is None and query_images is None and images is None: raise ValueError( """You have to specify at least one text or query image or image. All three cannot be none.""" ) if text is not None: if isinstance(lowercase , lowercase ) or (isinstance(lowercase , lowercase ) and not isinstance(text[0] , lowercase )): A_ : List[Any] = [self.tokenizer(lowercase , padding=lowercase , return_tensors=lowercase , **lowercase )] elif isinstance(lowercase , lowercase ) and isinstance(text[0] , lowercase ): A_ : Union[str, Any] = [] # Maximum number of queries across batch A_ : Dict = max([len(lowercase ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(lowercase ) != max_num_queries: A_ : Dict = t + [""" """] * (max_num_queries - len(lowercase )) A_ : Optional[Any] = self.tokenizer(lowercase , padding=lowercase , return_tensors=lowercase , **lowercase ) encodings.append(lowercase ) else: raise TypeError("""Input text should be a string, a list of strings or a nested list of strings""" ) if return_tensors == "np": A_ : List[Any] = np.concatenate([encoding["""input_ids"""] for encoding in encodings] , axis=0 ) A_ : Optional[Any] = np.concatenate([encoding["""attention_mask"""] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp A_ : Tuple = jnp.concatenate([encoding["""input_ids"""] for encoding in encodings] , axis=0 ) A_ : Any = jnp.concatenate([encoding["""attention_mask"""] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch A_ : List[str] = torch.cat([encoding["""input_ids"""] for encoding in encodings] , dim=0 ) A_ : Dict = torch.cat([encoding["""attention_mask"""] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf A_ : Tuple = tf.stack([encoding["""input_ids"""] for encoding in encodings] , axis=0 ) A_ : Dict = tf.stack([encoding["""attention_mask"""] for encoding in encodings] , axis=0 ) else: raise ValueError("""Target return tensor type could not be returned""" ) A_ : List[Any] = BatchEncoding() A_ : Optional[int] = input_ids A_ : Dict = attention_mask if query_images is not None: A_ : Optional[Any] = BatchEncoding() A_ : str = self.image_processor( lowercase , return_tensors=lowercase , **lowercase ).pixel_values A_ : List[str] = query_pixel_values if images is not None: A_ : Union[str, Any] = self.image_processor(lowercase , return_tensors=lowercase , **lowercase ) if text is not None and images is not None: A_ : Union[str, Any] = image_features.pixel_values return encoding elif query_images is not None and images is not None: A_ : List[str] = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**lowercase ) , tensor_type=lowercase ) def _a (self , *lowercase , **lowercase ): return self.image_processor.post_process(*lowercase , **lowercase ) def _a (self , *lowercase , **lowercase ): return self.image_processor.post_process_object_detection(*lowercase , **lowercase ) def _a (self , *lowercase , **lowercase ): return self.image_processor.post_process_image_guided_detection(*lowercase , **lowercase ) def _a (self , *lowercase , **lowercase ): return self.tokenizer.batch_decode(*lowercase , **lowercase ) def _a (self , *lowercase , **lowercase ): return self.tokenizer.decode(*lowercase , **lowercase ) @property def _a (self ): warnings.warn( """`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.""" , lowercase , ) return self.image_processor_class @property def _a (self ): warnings.warn( """`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.""" , lowercase , ) return self.image_processor
206
'''simple docstring''' from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline lowerCamelCase :Tuple = logging.get_logger(__name__) class _lowerCAmelCase ( __UpperCAmelCase ): def _a (self , lowercase ): if isinstance(lowercase , lowercase ): A_ : Tuple = [label.strip() for label in labels.split(""",""" ) if label.strip()] return labels def __call__(self , lowercase , lowercase , lowercase ): if len(lowercase ) == 0 or len(lowercase ) == 0: raise ValueError("""You must include at least one label and at least one sequence.""" ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( """The provided hypothesis_template \"{}\" was not able to be formatted with the target labels. """ """Make sure the passed template includes formatting syntax such as {{}} where the label should go.""" ).format(lowercase ) ) if isinstance(lowercase , lowercase ): A_ : Optional[Any] = [sequences] A_ : List[Any] = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(lowercase )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(__UpperCAmelCase ) class _lowerCAmelCase ( __UpperCAmelCase ): def __init__(self , lowercase=ZeroShotClassificationArgumentHandler() , *lowercase , **lowercase ): A_ : str = args_parser super().__init__(*lowercase , **lowercase ) if self.entailment_id == -1: logger.warning( """Failed to determine 'entailment' label id from the label2id mapping in the model config. Setting to """ """-1. Define a descriptive label2id mapping in the model config to ensure correct outputs.""" ) @property def _a (self ): for label, ind in self.model.config.labelaid.items(): if label.lower().startswith("""entail""" ): return ind return -1 def _a (self , lowercase , lowercase=True , lowercase=True , lowercase=TruncationStrategy.ONLY_FIRST , **lowercase ): A_ : Optional[int] = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( """Tokenizer was not supporting padding necessary for zero-shot, attempting to use """ """ `pad_token=eos_token`""" ) A_ : Any = self.tokenizer.eos_token try: A_ : Optional[Any] = self.tokenizer( lowercase , add_special_tokens=lowercase , return_tensors=lowercase , padding=lowercase , truncation=lowercase , ) except Exception as e: if "too short" in str(lowercase ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. A_ : str = self.tokenizer( lowercase , add_special_tokens=lowercase , return_tensors=lowercase , padding=lowercase , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def _a (self , **lowercase ): if kwargs.get("""multi_class""" , lowercase ) is not None: A_ : Union[str, Any] = kwargs["""multi_class"""] logger.warning( """The `multi_class` argument has been deprecated and renamed to `multi_label`. """ """`multi_class` will be removed in a future version of Transformers.""" ) A_ : Tuple = {} if "candidate_labels" in kwargs: A_ : Optional[int] = self._args_parser._parse_labels(kwargs["""candidate_labels"""] ) if "hypothesis_template" in kwargs: A_ : List[str] = kwargs["""hypothesis_template"""] A_ : Union[str, Any] = {} if "multi_label" in kwargs: A_ : Optional[int] = kwargs["""multi_label"""] return preprocess_params, {}, postprocess_params def __call__(self , lowercase , *lowercase , **lowercase , ): if len(lowercase ) == 0: pass elif len(lowercase ) == 1 and "candidate_labels" not in kwargs: A_ : Any = args[0] else: raise ValueError(F'Unable to understand extra arguments {args}' ) return super().__call__(lowercase , **lowercase ) def _a (self , lowercase , lowercase=None , lowercase="This example is {}." ): A_, A_ : Any = self._args_parser(lowercase , lowercase , lowercase ) for i, (candidate_label, sequence_pair) in enumerate(zip(lowercase , lowercase ) ): A_ : List[Any] = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(lowercase ) - 1, **model_input, } def _a (self , lowercase ): A_ : Dict = inputs["""candidate_label"""] A_ : Any = inputs["""sequence"""] A_ : List[Any] = {k: inputs[k] for k in self.tokenizer.model_input_names} A_ : Optional[int] = self.model(**lowercase ) A_ : Optional[Any] = { """candidate_label""": candidate_label, """sequence""": sequence, """is_last""": inputs["""is_last"""], **outputs, } return model_outputs def _a (self , lowercase , lowercase=False ): A_ : Tuple = [outputs["""candidate_label"""] for outputs in model_outputs] A_ : Optional[int] = [outputs["""sequence"""] for outputs in model_outputs] A_ : Union[str, Any] = np.concatenate([output["""logits"""].numpy() for output in model_outputs] ) A_ : List[str] = logits.shape[0] A_ : Optional[int] = len(lowercase ) A_ : int = N // n A_ : int = logits.reshape((num_sequences, n, -1) ) if multi_label or len(lowercase ) == 1: # softmax over the entailment vs. contradiction dim for each label independently A_ : Dict = self.entailment_id A_ : Dict = -1 if entailment_id == 0 else 0 A_ : str = reshaped_outputs[..., [contradiction_id, entailment_id]] A_ : Optional[Any] = np.exp(lowercase ) / np.exp(lowercase ).sum(-1 , keepdims=lowercase ) A_ : Tuple = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels A_ : Optional[int] = reshaped_outputs[..., self.entailment_id] A_ : Any = np.exp(lowercase ) / np.exp(lowercase ).sum(-1 , keepdims=lowercase ) A_ : List[str] = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
206
1
'''simple docstring''' def _lowerCamelCase ( lowercase : int ) -> int: _a = [1] _a , _a , _a = 0, 0, 0 _a = ugly_nums[ia] * 2 _a = ugly_nums[ia] * 3 _a = ugly_nums[ia] * 5 for _ in range(1 , lowercase ): _a = min(lowercase , lowercase , lowercase ) ugly_nums.append(lowercase ) if next_num == next_a: ia += 1 _a = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 _a = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 _a = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f"""{ugly_numbers(2_00) = }""")
346
'''simple docstring''' import re from filelock import FileLock try: import nltk lowerCAmelCase_ : Optional[int] = True except (ImportError, ModuleNotFoundError): lowerCAmelCase_ : Tuple = False if NLTK_AVAILABLE: with FileLock('.lock') as lock: nltk.download('punkt', quiet=True) def _lowerCamelCase ( lowercase : str ) -> str: re.sub("<n>" , "" , lowercase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(lowercase ) )
346
1
'''simple docstring''' import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) snake_case_ : str = logging.getLogger(__name__) def A__ ( ): _UpperCamelCase : List[Any] = argparse.ArgumentParser( description='Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids).' ) parser.add_argument('--file_path' , type=UpperCAmelCase_ , default='data/dump.txt' , help='The path to the data.' ) parser.add_argument('--tokenizer_type' , type=UpperCAmelCase_ , default='bert' , choices=['bert', 'roberta', 'gpt2'] ) parser.add_argument('--tokenizer_name' , type=UpperCAmelCase_ , default='bert-base-uncased' , help='The tokenizer to use.' ) parser.add_argument('--dump_file' , type=UpperCAmelCase_ , default='data/dump' , help='The dump file prefix.' ) _UpperCamelCase : Any = parser.parse_args() logger.info(f'Loading Tokenizer ({args.tokenizer_name})' ) if args.tokenizer_type == "bert": _UpperCamelCase : Optional[int] = BertTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[int] = tokenizer.special_tokens_map['cls_token'] # `[CLS]` _UpperCamelCase : Dict = tokenizer.special_tokens_map['sep_token'] # `[SEP]` elif args.tokenizer_type == "roberta": _UpperCamelCase : List[Any] = RobertaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Any = tokenizer.special_tokens_map['cls_token'] # `<s>` _UpperCamelCase : int = tokenizer.special_tokens_map['sep_token'] # `</s>` elif args.tokenizer_type == "gpt2": _UpperCamelCase : Optional[int] = GPTaTokenizer.from_pretrained(args.tokenizer_name ) _UpperCamelCase : Optional[Any] = tokenizer.special_tokens_map['bos_token'] # `<|endoftext|>` _UpperCamelCase : Any = tokenizer.special_tokens_map['eos_token'] # `<|endoftext|>` logger.info(f'Loading text from {args.file_path}' ) with open(args.file_path , 'r' , encoding='utf8' ) as fp: _UpperCamelCase : List[Any] = fp.readlines() logger.info('Start encoding' ) logger.info(f'{len(UpperCAmelCase_ )} examples to process.' ) _UpperCamelCase : int = [] _UpperCamelCase : Any = 0 _UpperCamelCase : Any = 1_0_0_0_0 _UpperCamelCase : Optional[Any] = time.time() for text in data: _UpperCamelCase : List[Any] = f'{bos} {text.strip()} {sep}' _UpperCamelCase : Any = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) rslt.append(UpperCAmelCase_ ) iter += 1 if iter % interval == 0: _UpperCamelCase : Union[str, Any] = time.time() logger.info(f'{iter} examples processed. - {(end-start):.2f}s/{interval}expl' ) _UpperCamelCase : Tuple = time.time() logger.info('Finished binarization' ) logger.info(f'{len(UpperCAmelCase_ )} examples processed.' ) _UpperCamelCase : Optional[int] = f'{args.dump_file}.{args.tokenizer_name}.pickle' _UpperCamelCase : List[str] = tokenizer.vocab_size if vocab_size < (1 << 1_6): _UpperCamelCase : List[Any] = [np.uintaa(UpperCAmelCase_ ) for d in rslt] else: _UpperCamelCase : Any = [np.intaa(UpperCAmelCase_ ) for d in rslt] random.shuffle(rslt_ ) logger.info(f'Dump to {dp_file}' ) with open(UpperCAmelCase_ , 'wb' ) as handle: pickle.dump(rslt_ , UpperCAmelCase_ , protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
83
import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def __init__( self,__lowerCamelCase,__lowerCamelCase=3,__lowerCamelCase=32,__lowerCamelCase=3,__lowerCamelCase=10,__lowerCamelCase=[10, 20, 30, 40],__lowerCamelCase=[1, 1, 2, 1],__lowerCamelCase=True,__lowerCamelCase=True,__lowerCamelCase="relu",__lowerCamelCase=3,__lowerCamelCase=None,): A__ = parent A__ = batch_size A__ = image_size A__ = num_channels A__ = embeddings_size A__ = hidden_sizes A__ = depths A__ = is_training A__ = use_labels A__ = hidden_act A__ = num_labels A__ = scope A__ = len(__lowerCamelCase ) def UpperCamelCase ( self ): A__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ = self.get_config() return config, pixel_values def UpperCamelCase ( self ): return RegNetConfig( num_channels=self.num_channels,embeddings_size=self.embeddings_size,hidden_sizes=self.hidden_sizes,depths=self.depths,hidden_act=self.hidden_act,num_labels=self.num_labels,image_size=self.image_size,) def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase ): A__ = FlaxRegNetModel(config=__lowerCamelCase ) A__ = model(__lowerCamelCase ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape,(self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32),) def UpperCamelCase ( self,__lowerCamelCase,__lowerCamelCase ): A__ = self.num_labels A__ = FlaxRegNetForImageClassification(config=__lowerCamelCase ) A__ = model(__lowerCamelCase ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.num_labels) ) def UpperCamelCase ( self ): A__ = self.prepare_config_and_inputs() A__ , A__ = config_and_inputs A__ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ , unittest.TestCase ): __SCREAMING_SNAKE_CASE = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def UpperCamelCase ( self ): A__ = FlaxRegNetModelTester(self ) A__ = ConfigTester(self,config_class=__lowerCamelCase,has_text_modality=__lowerCamelCase ) def UpperCamelCase ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ): return def UpperCamelCase ( self ): A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def UpperCamelCase ( self ): A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def UpperCamelCase ( self ): pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def UpperCamelCase ( self ): pass def UpperCamelCase ( self ): A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(__lowerCamelCase ) A__ = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ = [*signature.parameters.keys()] A__ = ['''pixel_values'''] self.assertListEqual(arg_names[:1],__lowerCamelCase ) def UpperCamelCase ( self ): def check_hidden_states_output(__lowerCamelCase,__lowerCamelCase,__lowerCamelCase ): A__ = model_class(__lowerCamelCase ) A__ = model(**self._prepare_for_class(__lowerCamelCase,__lowerCamelCase ) ) A__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A__ = self.model_tester.num_stages self.assertEqual(len(__lowerCamelCase ),expected_num_stages + 1 ) A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = True check_hidden_states_output(__lowerCamelCase,__lowerCamelCase,__lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ = True check_hidden_states_output(__lowerCamelCase,__lowerCamelCase,__lowerCamelCase ) def UpperCamelCase ( self ): A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): A__ = self._prepare_for_class(__lowerCamelCase,__lowerCamelCase ) A__ = model_class(__lowerCamelCase ) @jax.jit def model_jitted(__lowerCamelCase,**__lowerCamelCase ): return model(pixel_values=__lowerCamelCase,**__lowerCamelCase ) with self.subTest('''JIT Enabled''' ): A__ = model_jitted(**__lowerCamelCase ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): A__ = model_jitted(**__lowerCamelCase ).to_tuple() self.assertEqual(len(__lowerCamelCase ),len(__lowerCamelCase ) ) for jitted_output, output in zip(__lowerCamelCase,__lowerCamelCase ): self.assertEqual(jitted_output.shape,output.shape ) def UpperCamelCase__( )->Optional[int]: A__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_flax class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ): return AutoImageProcessor.from_pretrained('''facebook/regnet-y-040''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ): A__ = FlaxRegNetForImageClassification.from_pretrained('''facebook/regnet-y-040''' ) A__ = self.default_image_processor A__ = prepare_img() A__ = image_processor(images=__lowerCamelCase,return_tensors='''np''' ) A__ = model(**__lowerCamelCase ) # verify the logits A__ = (1, 1000) self.assertEqual(outputs.logits.shape,__lowerCamelCase ) A__ = jnp.array([-0.4180, -1.5051, -3.4836] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3],__lowerCamelCase,atol=1E-4 ) )
193
0
import collections import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __A = logging.get_logger(__name__) __A = '▁' __A = {'vocab_file': 'prophetnet.tokenizer'} __A = { 'vocab_file': { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/prophetnet.tokenizer' ), } } __A = { 'microsoft/xprophetnet-large-wiki100-cased': {'do_lower_case': False}, } __A = { 'microsoft/xprophetnet-large-wiki100-cased': 5_12, } def lowerCamelCase_ ( UpperCamelCase__ : Optional[Any] ): """simple docstring""" __lowerCamelCase = collections.OrderedDict() with open(_UpperCAmelCase , 'r' , encoding='utf-8' ) as reader: __lowerCamelCase = reader.readlines() for index, token in enumerate(_UpperCAmelCase ): __lowerCamelCase = token.rstrip('\n' ) __lowerCamelCase = index return vocab class __lowerCAmelCase ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCamelCase__ , lowerCamelCase__="[SEP]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[UNK]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__ = None , **lowerCamelCase__ , ) -> None: '''simple docstring''' __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase__ , ) try: import sentencepiece as spm except ImportError: logger.warning( 'You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece' ' pip install sentencepiece' ) raise __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase__ ) ) __lowerCamelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # put special tokens and [unused] tokens into the vocab __lowerCamelCase = {'[PAD]': 0, '[CLS]': 1, '[SEP]': 2, '[UNK]': 3, '[MASK]': 4} for i in range(10 ): __lowerCamelCase = f"""[unused{i}]""" __lowerCamelCase = 5 + i # The first "real" token "," has position 15 in the embedding vocab and position 3 in the spm vocab __lowerCamelCase = 12 __lowerCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} for k in self.fairseq_tokens_to_ids.keys(): self.unique_no_split_tokens.append(lowerCamelCase__ ) def __getstate__( self ) -> List[str]: '''simple docstring''' __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self , lowerCamelCase__ ) -> Dict: '''simple docstring''' __lowerCamelCase = d try: import sentencepiece as spm except ImportError: logger.warning( 'You need to install SentencePiece to use XLMRobertaTokenizer: https://github.com/google/sentencepiece' ' pip install sentencepiece' ) raise # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return ([0] * len(lowerCamelCase__ )) + [1] return ([0] * len(lowerCamelCase__ )) + [1] + ([0] * len(lowerCamelCase__ )) + [1] def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> List[int]: '''simple docstring''' __lowerCamelCase = [self.sep_token_id] if token_ids_a is None: return len(token_ids_a + sep ) * [0] return len(token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def lowercase_ ( self ) -> Union[str, Any]: '''simple docstring''' return len(self.sp_model ) + self.fairseq_offset def lowercase_ ( self ) -> int: '''simple docstring''' __lowerCamelCase = {self.convert_ids_to_tokens(lowerCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase_ ( self , lowerCamelCase__ ) -> str: '''simple docstring''' return self.sp_model.encode(lowerCamelCase__ , out_type=lowerCamelCase__ ) def lowercase_ ( self , lowerCamelCase__ ) -> Union[str, Any]: '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __lowerCamelCase = self.sp_model.PieceToId(lowerCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowercase_ ( self , lowerCamelCase__ ) -> Union[str, Any]: '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowercase_ ( self , lowerCamelCase__ ) -> str: '''simple docstring''' __lowerCamelCase = ''.join(lowerCamelCase__ ).replace(lowerCamelCase__ , ' ' ).strip() return out_string def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase__ , 'wb' ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase__ ) return (out_vocab_file,) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__ = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.sep_token_id] __lowerCamelCase = [self.sep_token_id] return token_ids_a + sep + token_ids_a + sep
359
import requests __A = "https://newsapi.org/v1/articles?source=bbc-news&sortBy=top&apiKey=" def lowerCamelCase_ ( UpperCamelCase__ : str ) -> None: """simple docstring""" __lowerCamelCase = requests.get(_NEWS_API + bbc_news_api_key ).json() # each article in the list is a dict for i, article in enumerate(bbc_news_page['articles'] , 1 ): print(F"""{i}.) {article['title']}""" ) if __name__ == "__main__": fetch_bbc_news(bbc_news_api_key="<Your BBC News API key goes here>")
348
0
import itertools import math def _snake_case( SCREAMING_SNAKE_CASE__ ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(SCREAMING_SNAKE_CASE__ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def _snake_case( ) -> List[str]: lowercase : List[str] = 2 while True: if is_prime(SCREAMING_SNAKE_CASE__ ): yield num num += 1 def _snake_case( SCREAMING_SNAKE_CASE__ = 10_001 ) -> int: return next(itertools.islice(prime_generator() , nth - 1 , SCREAMING_SNAKE_CASE__ ) ) if __name__ == "__main__": print(F'''{solution() = }''')
20
'''simple docstring''' from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) @add_end_docstrings(A__ ) class __UpperCAmelCase ( A__ ): '''simple docstring''' def __init__(self : Tuple , *_lowerCAmelCase : List[str] , **_lowerCAmelCase : List[str] ): super().__init__(*_lowerCAmelCase , **_lowerCAmelCase ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def A (self : Any , _lowerCAmelCase : str=None ): A = {} if top_k is not None: A = top_k return {}, {}, postprocess_params def __call__(self : str , _lowerCAmelCase : Union[str, List[str], "Image.Image", List["Image.Image"]] , **_lowerCAmelCase : int ): return super().__call__(_lowerCAmelCase , **_lowerCAmelCase ) def A (self : List[str] , _lowerCAmelCase : List[Any] ): A = load_image(_lowerCAmelCase ) A = self.image_processor(images=_lowerCAmelCase , return_tensors=self.framework ) return model_inputs def A (self : Union[str, Any] , _lowerCAmelCase : Optional[int] ): A = self.model(**_lowerCAmelCase ) return model_outputs def A (self : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : int=5 ): if top_k > self.model.config.num_labels: A = self.model.config.num_labels if self.framework == "pt": A = model_outputs.logits.softmax(-1 )[0] A , A = probs.topk(_lowerCAmelCase ) elif self.framework == "tf": A = stable_softmax(model_outputs.logits , axis=-1 )[0] A = tf.math.top_k(_lowerCAmelCase , k=_lowerCAmelCase ) A , A = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(F"""Unsupported framework: {self.framework}""" ) A = scores.tolist() A = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(_lowerCAmelCase , _lowerCAmelCase )]
258
0
'''simple docstring''' import copy import os import cva import numpy as np from matplotlib import pyplot as plt class __snake_case: '''simple docstring''' def __init__( self ) -> Optional[int]: lowerCAmelCase = """""" lowerCAmelCase = """""" lowerCAmelCase = [] lowerCAmelCase = 0 lowerCAmelCase = 256 lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = 0 lowerCAmelCase = 0 def __snake_case ( self , A_ ) -> Dict: lowerCAmelCase = cva.imread(A_ , 0 ) lowerCAmelCase = copy.deepcopy(self.img ) lowerCAmelCase, lowerCAmelCase, lowerCAmelCase = plt.hist(self.img.ravel() , 256 , [0, 256] , label="""x""" ) lowerCAmelCase = np.sum(A_ ) for i in range(len(A_ ) ): lowerCAmelCase = x[i] / self.k self.sk += prk lowerCAmelCase = (self.L - 1) * self.sk if self.rem != 0: lowerCAmelCase = int(last % last ) lowerCAmelCase = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(A_ ) lowerCAmelCase = int(np.ma.count(self.img ) / self.img[1].size ) lowerCAmelCase = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): lowerCAmelCase = self.img[j][i] if num != self.last_list[num]: lowerCAmelCase = self.last_list[num] cva.imwrite("""output_data/output.jpg""" , self.img ) def __snake_case ( self ) -> Tuple: plt.hist(self.img.ravel() , 256 , [0, 256] ) def __snake_case ( self ) -> Dict: cva.imshow("""Output-Image""" , self.img ) cva.imshow("""Input-Image""" , self.original_image ) cva.waitKey(5000 ) cva.destroyAllWindows() if __name__ == "__main__": UpperCAmelCase = os.path.join(os.path.basename(__file__), 'image_data/input.jpg') UpperCAmelCase = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
362
'''simple docstring''' class __snake_case( _lowerCAmelCase ): '''simple docstring''' pass class __snake_case( _lowerCAmelCase ): '''simple docstring''' pass class __snake_case: '''simple docstring''' def __init__( self ) -> int: lowerCAmelCase = [ [], [], [], ] def __snake_case ( self , A_ , A_ ) -> None: try: if len(self.queues[priority] ) >= 100: raise OverflowError("""Maximum queue size is 100""" ) self.queues[priority].append(A_ ) except IndexError: raise ValueError("""Valid priorities are 0, 1, and 2""" ) def __snake_case ( self ) -> int: for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("""All queues are empty""" ) def __str__( self ) -> str: return "\n".join(f'Priority {i}: {q}' for i, q in enumerate(self.queues ) ) class __snake_case: '''simple docstring''' def __init__( self ) -> Dict: lowerCAmelCase = [] def __snake_case ( self , A_ ) -> None: if len(self.queue ) == 100: raise OverFlowError("""Maximum queue size is 100""" ) self.queue.append(A_ ) def __snake_case ( self ) -> int: if not self.queue: raise UnderFlowError("""The queue is empty""" ) else: lowerCAmelCase = min(self.queue ) self.queue.remove(A_ ) return data def __str__( self ) -> str: return str(self.queue ) def _snake_case ( ) -> Tuple: """simple docstring""" lowerCAmelCase = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 100 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 128 ) print(_SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(_SCREAMING_SNAKE_CASE ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def _snake_case ( ) -> Optional[Any]: """simple docstring""" lowerCAmelCase = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(100 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(128 ) print(_SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(_SCREAMING_SNAKE_CASE ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
187
0
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class UpperCAmelCase_ ( unittest.TestCase): lowerCamelCase__ : Union[str, Any] = ViTImageProcessor if is_vision_available() else None @property def _UpperCAmelCase ( self ) -> Optional[Any]: return self.image_processor_tester.prepare_image_processor_dict() def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Union[str, Any] = (3, 3_2, 1_2_8) lowercase__ : Dict = tempfile.mkdtemp() # fmt: off lowercase__ : Tuple = ['[GO]', '[s]', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z'] # fmt: on lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Optional[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(a ) + '\n' ) lowercase__ : Dict = { 'do_normalize': False, 'do_resize': True, 'image_processor_type': 'ViTImageProcessor', 'resample': 3, 'size': {'height': 3_2, 'width': 1_2_8}, } lowercase__ : Dict = os.path.join(self.tmpdirname , a ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(a , a ) def _UpperCAmelCase ( self , **a ) -> Union[str, Any]: return MgpstrTokenizer.from_pretrained(self.tmpdirname , **a ) def _UpperCAmelCase ( self , **a ) -> Any: return ViTImageProcessor.from_pretrained(self.tmpdirname , **a ) def _UpperCAmelCase ( self ) -> List[str]: shutil.rmtree(self.tmpdirname ) def _UpperCAmelCase ( self ) -> str: lowercase__ : List[str] = np.random.randint(2_5_5 , size=(3, 3_0, 4_0_0) , dtype=np.uinta ) lowercase__ : Union[str, Any] = Image.fromarray(np.moveaxis(a , 0 , -1 ) ) return image_input def _UpperCAmelCase ( self ) -> int: lowercase__ : Optional[int] = self.get_tokenizer() lowercase__ : Optional[Any] = self.get_image_processor() lowercase__ : Optional[Any] = MgpstrProcessor(tokenizer=a , image_processor=a ) processor.save_pretrained(self.tmpdirname ) lowercase__ : Dict = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=a ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , a ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , a ) def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Tuple = self.get_tokenizer() lowercase__ : Dict = self.get_image_processor() lowercase__ : Optional[Any] = MgpstrProcessor(tokenizer=a , image_processor=a ) processor.save_pretrained(self.tmpdirname ) lowercase__ : List[str] = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) lowercase__ : Optional[int] = self.get_image_processor(do_normalize=a , padding_value=1.0 ) lowercase__ : int = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=a , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , a ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Tuple = self.get_image_processor() lowercase__ : Tuple = self.get_tokenizer() lowercase__ : Tuple = MgpstrProcessor(tokenizer=a , image_processor=a ) lowercase__ : str = self.prepare_image_inputs() lowercase__ : List[str] = image_processor(a , return_tensors='np' ) lowercase__ : str = processor(images=a , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Dict = self.get_image_processor() lowercase__ : Any = self.get_tokenizer() lowercase__ : Union[str, Any] = MgpstrProcessor(tokenizer=a , image_processor=a ) lowercase__ : List[str] = 'test' lowercase__ : str = processor(text=a ) lowercase__ : List[Any] = tokenizer(a ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Any = self.get_image_processor() lowercase__ : List[Any] = self.get_tokenizer() lowercase__ : int = MgpstrProcessor(tokenizer=a , image_processor=a ) lowercase__ : Union[str, Any] = 'test' lowercase__ : List[Any] = self.prepare_image_inputs() lowercase__ : List[Any] = processor(text=a , images=a ) self.assertListEqual(list(inputs.keys() ) , ['pixel_values', 'labels'] ) # test if it raises when no input is passed with pytest.raises(a ): processor() def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : List[str] = self.get_image_processor() lowercase__ : Dict = self.get_tokenizer() lowercase__ : List[str] = MgpstrProcessor(tokenizer=a , image_processor=a ) lowercase__ : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] lowercase__ : List[str] = processor.char_decode(a ) lowercase__ : Union[str, Any] = tokenizer.batch_decode(a ) lowercase__ : Any = [seq.replace(' ' , '' ) for seq in decoded_tok] self.assertListEqual(a , a ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : List[Any] = self.get_image_processor() lowercase__ : Tuple = self.get_tokenizer() lowercase__ : Dict = MgpstrProcessor(tokenizer=a , image_processor=a ) lowercase__ : Optional[Any] = None lowercase__ : int = self.prepare_image_inputs() lowercase__ : Optional[int] = processor(text=a , images=a ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[str] = self.get_image_processor() lowercase__ : Any = self.get_tokenizer() lowercase__ : str = MgpstrProcessor(tokenizer=a , image_processor=a ) lowercase__ : List[Any] = torch.randn(1 , 2_7 , 3_8 ) lowercase__ : List[str] = torch.randn(1 , 2_7 , 5_0_2_5_7 ) lowercase__ : List[str] = torch.randn(1 , 2_7 , 3_0_5_2_2 ) lowercase__ : str = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['generated_text', 'scores', 'char_preds', 'bpe_preds', 'wp_preds'] )
77
'''simple docstring''' from typing import Any def UpperCamelCase_ ( _UpperCAmelCase : list , _UpperCAmelCase : list , _UpperCAmelCase : dict , _UpperCAmelCase : dict , _UpperCAmelCase : dict , ) -> list: """simple docstring""" _validation( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) # Creates data structures and fill initial step _UpperCAmelCase : dict = {} _UpperCAmelCase : dict = {} for state in states_space: _UpperCAmelCase : Union[str, Any] = observations_space[0] _UpperCAmelCase : Tuple = ( initial_probabilities[state] * emission_probabilities[state][observation] ) _UpperCAmelCase : List[str] = None # Fills the data structure with the probabilities of # different transitions and pointers to previous states for o in range(1 , len(_UpperCAmelCase ) ): _UpperCAmelCase : Optional[Any] = observations_space[o] _UpperCAmelCase : int = observations_space[o - 1] for state in states_space: # Calculates the argmax for probability function _UpperCAmelCase : str = "" _UpperCAmelCase : Tuple = -1 for k_state in states_space: _UpperCAmelCase : Any = ( probabilities[(k_state, prior_observation)] * transition_probabilities[k_state][state] * emission_probabilities[state][observation] ) if probability > max_probability: _UpperCAmelCase : Union[str, Any] = probability _UpperCAmelCase : str = k_state # Update probabilities and pointers dicts _UpperCAmelCase : Optional[int] = ( probabilities[(arg_max, prior_observation)] * transition_probabilities[arg_max][state] * emission_probabilities[state][observation] ) _UpperCAmelCase : Tuple = arg_max # The final observation _UpperCAmelCase : Optional[Any] = observations_space[len(_UpperCAmelCase ) - 1] # argmax for given final observation _UpperCAmelCase : List[str] = "" _UpperCAmelCase : Any = -1 for k_state in states_space: _UpperCAmelCase : Optional[int] = probabilities[(k_state, final_observation)] if probability > max_probability: _UpperCAmelCase : int = probability _UpperCAmelCase : Dict = k_state _UpperCAmelCase : Dict = arg_max # Process pointers backwards _UpperCAmelCase : List[Any] = last_state _UpperCAmelCase : str = [] for o in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): result.append(_UpperCAmelCase ) _UpperCAmelCase : List[Any] = pointers[previous, observations_space[o]] result.reverse() return result def UpperCamelCase_ ( _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Any , ) -> None: """simple docstring""" _validate_not_empty( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) _validate_lists(_UpperCAmelCase , _UpperCAmelCase ) _validate_dicts( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase_ ( _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Any , ) -> None: """simple docstring""" if not all( [ observations_space, states_space, initial_probabilities, transition_probabilities, emission_probabilities, ] ): raise ValueError("There's an empty parameter" ) def UpperCamelCase_ ( _UpperCAmelCase : Any , _UpperCAmelCase : Any ) -> None: """simple docstring""" _validate_list(_UpperCAmelCase , "observations_space" ) _validate_list(_UpperCAmelCase , "states_space" ) def UpperCamelCase_ ( _UpperCAmelCase : Any , _UpperCAmelCase : str ) -> None: """simple docstring""" if not isinstance(_object , _UpperCAmelCase ): _UpperCAmelCase : Optional[int] = F"""{var_name} must be a list""" raise ValueError(_UpperCAmelCase ) else: for x in _object: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): _UpperCAmelCase : Optional[int] = F"""{var_name} must be a list of strings""" raise ValueError(_UpperCAmelCase ) def UpperCamelCase_ ( _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Any , ) -> None: """simple docstring""" _validate_dict(_UpperCAmelCase , "initial_probabilities" , _UpperCAmelCase ) _validate_nested_dict(_UpperCAmelCase , "transition_probabilities" ) _validate_nested_dict(_UpperCAmelCase , "emission_probabilities" ) def UpperCamelCase_ ( _UpperCAmelCase : Any , _UpperCAmelCase : str ) -> None: """simple docstring""" _validate_dict(_object , _UpperCAmelCase , _UpperCAmelCase ) for x in _object.values(): _validate_dict(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase_ ( _UpperCAmelCase : Any , _UpperCAmelCase : str , _UpperCAmelCase : type , _UpperCAmelCase : bool = False ) -> None: """simple docstring""" if not isinstance(_object , _UpperCAmelCase ): _UpperCAmelCase : Any = F"""{var_name} must be a dict""" raise ValueError(_UpperCAmelCase ) if not all(isinstance(_UpperCAmelCase , _UpperCAmelCase ) for x in _object ): _UpperCAmelCase : Tuple = F"""{var_name} all keys must be strings""" raise ValueError(_UpperCAmelCase ) if not all(isinstance(_UpperCAmelCase , _UpperCAmelCase ) for x in _object.values() ): _UpperCAmelCase : List[str] = "nested dictionary " if nested else "" _UpperCAmelCase : List[str] = F"""{var_name} {nested_text}all values must be {value_type.__name__}""" raise ValueError(_UpperCAmelCase ) if __name__ == "__main__": from doctest import testmod testmod()
31
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available snake_case_ : Any = { 'configuration_mvp': ['MVP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MvpConfig', 'MvpOnnxConfig'], 'tokenization_mvp': ['MvpTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : Dict = ['MvpTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ : List[Any] = [ 'MVP_PRETRAINED_MODEL_ARCHIVE_LIST', 'MvpForCausalLM', 'MvpForConditionalGeneration', 'MvpForQuestionAnswering', 'MvpForSequenceClassification', 'MvpModel', 'MvpPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys snake_case_ : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
236
'''simple docstring''' import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging snake_case_ : List[Any] = logging.get_logger(__name__) snake_case_ : str = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt'} # See all BART models at https://huggingface.co/models?filter=bart snake_case_ : Union[str, Any] = { 'vocab_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/vocab.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/vocab.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json', }, 'merges_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/merges.txt', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/merges.txt', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt', }, } snake_case_ : Any = { 'facebook/bart-base': 1024, 'facebook/bart-large': 1024, 'facebook/bart-large-mnli': 1024, 'facebook/bart-large-cnn': 1024, 'facebook/bart-large-xsum': 1024, 'yjernite/bart_eli5': 1024, } @lru_cache() def A__ ( ): _UpperCamelCase : str = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) _UpperCamelCase : Any = bs[:] _UpperCamelCase : Union[str, Any] = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCAmelCase_ ) cs.append(2**8 + n ) n += 1 _UpperCamelCase : Any = [chr(UpperCAmelCase_ ) for n in cs] return dict(zip(UpperCAmelCase_ , UpperCAmelCase_ ) ) def A__ ( UpperCAmelCase_ ): _UpperCamelCase : Tuple = set() _UpperCamelCase : Dict = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _UpperCamelCase : Any = char return pairs class lowercase__ ( lowercase ): lowercase__ = VOCAB_FILES_NAMES lowercase__ = PRETRAINED_VOCAB_FILES_MAP lowercase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ = ["""input_ids""", """attention_mask"""] def __init__( self : Optional[int] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[Any]="replace" ,lowerCamelCase__ : str="<s>" ,lowerCamelCase__ : str="</s>" ,lowerCamelCase__ : str="</s>" ,lowerCamelCase__ : Any="<s>" ,lowerCamelCase__ : Tuple="<unk>" ,lowerCamelCase__ : List[str]="<pad>" ,lowerCamelCase__ : Optional[Any]="<mask>" ,lowerCamelCase__ : Tuple=False ,**lowerCamelCase__ : List[str] ,): '''simple docstring''' _UpperCamelCase : Dict = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else bos_token _UpperCamelCase : Tuple = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else eos_token _UpperCamelCase : Optional[int] = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else sep_token _UpperCamelCase : Tuple = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else cls_token _UpperCamelCase : Union[str, Any] = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else unk_token _UpperCamelCase : List[str] = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _UpperCamelCase : str = AddedToken(lowerCamelCase__ ,lstrip=lowerCamelCase__ ,rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ ,lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ ,bos_token=lowerCamelCase__ ,eos_token=lowerCamelCase__ ,unk_token=lowerCamelCase__ ,sep_token=lowerCamelCase__ ,cls_token=lowerCamelCase__ ,pad_token=lowerCamelCase__ ,mask_token=lowerCamelCase__ ,add_prefix_space=lowerCamelCase__ ,**lowerCamelCase__ ,) with open(lowerCamelCase__ ,encoding='utf-8' ) as vocab_handle: _UpperCamelCase : List[str] = json.load(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = {v: k for k, v in self.encoder.items()} _UpperCamelCase : Optional[Any] = errors # how to handle errors in decoding _UpperCamelCase : Tuple = bytes_to_unicode() _UpperCamelCase : List[str] = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ ,encoding='utf-8' ) as merges_handle: _UpperCamelCase : Dict = merges_handle.read().split('\n' )[1:-1] _UpperCamelCase : str = [tuple(merge.split() ) for merge in bpe_merges] _UpperCamelCase : Dict = dict(zip(lowerCamelCase__ ,range(len(lowerCamelCase__ ) ) ) ) _UpperCamelCase : Tuple = {} _UpperCamelCase : List[Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _UpperCamelCase : Any = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' return len(self.encoder ) def UpperCamelCase_ ( self : str ): '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Any ): '''simple docstring''' if token in self.cache: return self.cache[token] _UpperCamelCase : Dict = tuple(lowerCamelCase__ ) _UpperCamelCase : List[str] = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _UpperCamelCase : List[str] = min(lowerCamelCase__ ,key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ ,float('inf' ) ) ) if bigram not in self.bpe_ranks: break _UpperCamelCase , _UpperCamelCase : int = bigram _UpperCamelCase : Optional[int] = [] _UpperCamelCase : Dict = 0 while i < len(lowerCamelCase__ ): try: _UpperCamelCase : int = word.index(lowerCamelCase__ ,lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _UpperCamelCase : Dict = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _UpperCamelCase : int = tuple(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = new_word if len(lowerCamelCase__ ) == 1: break else: _UpperCamelCase : Any = get_pairs(lowerCamelCase__ ) _UpperCamelCase : int = ' '.join(lowerCamelCase__ ) _UpperCamelCase : List[Any] = word return word def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : Optional[Any] ): '''simple docstring''' _UpperCamelCase : int = [] for token in re.findall(self.pat ,lowerCamelCase__ ): _UpperCamelCase : int = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(' ' ) ) return bpe_tokens def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : List[Any] ): '''simple docstring''' return self.encoder.get(lowerCamelCase__ ,self.encoder.get(self.unk_token ) ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : int ): '''simple docstring''' return self.decoder.get(lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ,lowerCamelCase__ : Any ): '''simple docstring''' _UpperCamelCase : Dict = ''.join(lowerCamelCase__ ) _UpperCamelCase : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' ,errors=self.errors ) return text def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(lowerCamelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCamelCase : List[Any] = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _UpperCamelCase : Union[str, Any] = os.path.join( lowerCamelCase__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(lowerCamelCase__ ,'w' ,encoding='utf-8' ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCamelCase__ ,ensure_ascii=lowerCamelCase__ ) + '\n' ) _UpperCamelCase : Optional[Any] = 0 with open(lowerCamelCase__ ,'w' ,encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' ' Please check that the tokenizer is not corrupted!' ) _UpperCamelCase : int = token_index writer.write(' '.join(lowerCamelCase__ ) + '\n' ) index += 1 return vocab_file, merge_file def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _UpperCamelCase : Optional[Any] = [self.cls_token_id] _UpperCamelCase : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ,lowerCamelCase__ : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ ,token_ids_a=lowerCamelCase__ ,already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def UpperCamelCase_ ( self : Tuple ,lowerCamelCase__ : List[int] ,lowerCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' _UpperCamelCase : Tuple = [self.sep_token_id] _UpperCamelCase : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Tuple=False ,**lowerCamelCase__ : Optional[int] ): '''simple docstring''' _UpperCamelCase : Tuple = kwargs.pop('add_prefix_space' ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _UpperCamelCase : List[str] = ' ' + text return (text, kwargs)
236
1
from math import pi, sqrt def _snake_case( SCREAMING_SNAKE_CASE__ ) -> float: if num <= 0: raise ValueError("""math domain error""" ) if num > 171.5: raise OverflowError("""math range error""" ) elif num - int(SCREAMING_SNAKE_CASE__ ) not in (0, 0.5): raise NotImplementedError("""num must be an integer or a half-integer""" ) elif num == 0.5: return sqrt(SCREAMING_SNAKE_CASE__ ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def _snake_case( ) -> None: assert gamma(0.5 ) == sqrt(SCREAMING_SNAKE_CASE__ ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() lowercase : Optional[Any] = 1.0 while num: lowercase : Dict = float(input("""Gamma of: """)) print(F'''gamma({num}) = {gamma(num)}''') print("""\nEnter 0 to exit...""")
20
from ...configuration_utils import PretrainedConfig from ...utils import logging _SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) _SCREAMING_SNAKE_CASE : Union[str, Any] = { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json''', } class UpperCAmelCase__ ( A__ ): """simple docstring""" a = "lxmert" a = {} def __init__( self : Union[str, Any] , __lowerCamelCase : List[str]=3_0522 , __lowerCamelCase : Union[str, Any]=768 , __lowerCamelCase : Dict=12 , __lowerCamelCase : Union[str, Any]=9500 , __lowerCamelCase : Union[str, Any]=1600 , __lowerCamelCase : Any=400 , __lowerCamelCase : List[str]=3072 , __lowerCamelCase : List[str]="gelu" , __lowerCamelCase : Optional[Any]=0.1 , __lowerCamelCase : Union[str, Any]=0.1 , __lowerCamelCase : Optional[Any]=512 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Any=0.02 , __lowerCamelCase : Any=1e-12 , __lowerCamelCase : List[Any]=9 , __lowerCamelCase : Any=5 , __lowerCamelCase : List[str]=5 , __lowerCamelCase : Optional[Any]=2048 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : List[str]=6.67 , __lowerCamelCase : Dict=True , __lowerCamelCase : Any=True , __lowerCamelCase : Any=True , __lowerCamelCase : Tuple=True , __lowerCamelCase : Optional[Any]=True , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Any=True , **__lowerCamelCase : Optional[Any] , ) -> Any: SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = layer_norm_eps SCREAMING_SNAKE_CASE__ = num_qa_labels SCREAMING_SNAKE_CASE__ = num_object_labels SCREAMING_SNAKE_CASE__ = num_attr_labels SCREAMING_SNAKE_CASE__ = l_layers SCREAMING_SNAKE_CASE__ = x_layers SCREAMING_SNAKE_CASE__ = r_layers SCREAMING_SNAKE_CASE__ = visual_feat_dim SCREAMING_SNAKE_CASE__ = visual_pos_dim SCREAMING_SNAKE_CASE__ = visual_loss_normalizer SCREAMING_SNAKE_CASE__ = task_matched SCREAMING_SNAKE_CASE__ = task_mask_lm SCREAMING_SNAKE_CASE__ = task_obj_predict SCREAMING_SNAKE_CASE__ = task_qa SCREAMING_SNAKE_CASE__ = visual_obj_loss SCREAMING_SNAKE_CASE__ = visual_attr_loss SCREAMING_SNAKE_CASE__ = visual_feat_loss SCREAMING_SNAKE_CASE__ = {'''vision''': r_layers, '''cross_encoder''': x_layers, '''language''': l_layers} super().__init__(**__lowerCamelCase )
314
0
from math import factorial, radians def _a ( SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : int = 18 , SCREAMING_SNAKE_CASE_ : int = 10 ): __lowerCAmelCase = angle_in_degrees - ((angle_in_degrees // 3_60.0) * 3_60.0) # Converting from degrees to radians __lowerCAmelCase = radians(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = angle_in_radians __lowerCAmelCase = 3 __lowerCAmelCase = -1 for _ in range(SCREAMING_SNAKE_CASE_ ): result += (b * (angle_in_radians**a)) / factorial(SCREAMING_SNAKE_CASE_ ) __lowerCAmelCase = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": __import__("""doctest""").testmod()
358
from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class a__ ( snake_case__ ): _a : Optional[int] = """new-model""" if is_tf_available(): class a__ ( snake_case__ ): _a : Dict = NewModelConfig @require_tf class a__ ( unittest.TestCase ): @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = "bert-base-cased" __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModel.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = "bert-base-cased" __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelForPreTraining.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelForCausalLM.from_pretrained(_A ) __lowerCAmelCase , __lowerCAmelCase = TFAutoModelForCausalLM.from_pretrained(_A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelWithLMHead.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelForMaskedLM.from_pretrained(_A ) __lowerCAmelCase , __lowerCAmelCase = TFAutoModelForMaskedLM.from_pretrained(_A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(_A ) __lowerCAmelCase , __lowerCAmelCase = TFAutoModelForSeqaSeqLM.from_pretrained(_A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelForSequenceClassification.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelForQuestionAnswering.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow @require_tensorflow_probability def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: __lowerCAmelCase = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = TFAutoModelForTableQuestionAnswering.from_pretrained(_A ) __lowerCAmelCase , __lowerCAmelCase = TFAutoModelForTableQuestionAnswering.from_pretrained( _A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFAutoModelWithLMHead.from_pretrained(_A ) self.assertIsInstance(_A , _A ) self.assertEqual(model.num_parameters() , 1_4_4_1_0 ) self.assertEqual(model.num_parameters(only_trainable=_A ) , 1_4_4_1_0 ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFAutoModelWithLMHead.from_pretrained(_A ) self.assertIsInstance(_A , _A ) self.assertEqual(model.num_parameters() , 1_4_4_1_0 ) self.assertEqual(model.num_parameters(only_trainable=_A ) , 1_4_4_1_0 ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFAutoModel.from_pretrained("sgugger/funnel-random-tiny" ) self.assertIsInstance(_A , _A ) __lowerCAmelCase = copy.deepcopy(model.config ) __lowerCAmelCase = ["FunnelBaseModel"] __lowerCAmelCase = TFAutoModel.from_config(_A ) self.assertIsInstance(_A , _A ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(_A ) __lowerCAmelCase = TFAutoModel.from_pretrained(_A ) self.assertIsInstance(_A , _A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" try: AutoConfig.register("new-model" , _A ) __lowerCAmelCase = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(_A ): auto_class.register(_A , _A ) auto_class.register(_A , _A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_A ): auto_class.register(_A , _A ) # Now that the config is registered, it can be used as any other config with the auto-API __lowerCAmelCase = BertModelTester(self ).get_config() __lowerCAmelCase = NewModelConfig(**tiny_config.to_dict() ) __lowerCAmelCase = auto_class.from_config(_A ) self.assertIsInstance(_A , _A ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(_A ) __lowerCAmelCase = auto_class.from_pretrained(_A ) self.assertIsInstance(_A , _A ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaisesRegex( _A , "bert-base is not a local folder and is not a valid model identifier" ): __lowerCAmelCase = TFAutoModel.from_pretrained("bert-base" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaisesRegex( _A , R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __lowerCAmelCase = TFAutoModel.from_pretrained(_A , revision="aaaaaa" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaisesRegex( _A , "hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin" , ): __lowerCAmelCase = TFAutoModel.from_pretrained("hf-internal-testing/config-no-model" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" with self.assertRaisesRegex(_A , "Use `from_pt=True` to load this model" ): __lowerCAmelCase = TFAutoModel.from_pretrained("hf-internal-testing/tiny-bert-pt-only" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" __lowerCAmelCase = TFAutoModel.from_pretrained("hf-internal-testing/tiny-random-bert" ) with RequestCounter() as counter: __lowerCAmelCase = TFAutoModel.from_pretrained("hf-internal-testing/tiny-random-bert" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint __lowerCAmelCase = TFAutoModel.from_pretrained("ArthurZ/tiny-random-bert-sharded" ) with RequestCounter() as counter: __lowerCAmelCase = TFAutoModel.from_pretrained("ArthurZ/tiny-random-bert-sharded" ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
102
0
import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) __a = pytest.mark.integration @pytest.mark.parametrize('''path''' , ['''paws''', '''csv'''] ) def a ( snake_case__: str , snake_case__: Tuple ): '''simple docstring''' inspect_dataset(snake_case__ , snake_case__ ) lowercase_ = path + '''.py''' assert script_name in os.listdir(snake_case__ ) assert "__pycache__" not in os.listdir(snake_case__ ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' , ['''accuracy'''] ) def a ( snake_case__: List[str] , snake_case__: Dict ): '''simple docstring''' inspect_metric(snake_case__ , snake_case__ ) lowercase_ = path + '''.py''' assert script_name in os.listdir(snake_case__ ) assert "__pycache__" not in os.listdir(snake_case__ ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def a ( snake_case__: Dict , snake_case__: Optional[int] , snake_case__: Dict ): '''simple docstring''' lowercase_ = get_dataset_config_info(snake_case__ , config_name=snake_case__ ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def a ( snake_case__: Any , snake_case__: Optional[int] , snake_case__: List[Any] ): '''simple docstring''' with pytest.raises(snake_case__ ): get_dataset_config_info(snake_case__ , config_name=snake_case__ ) @pytest.mark.parametrize( '''path, expected''' , [ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] , ) def a ( snake_case__: int , snake_case__: Optional[int] ): '''simple docstring''' lowercase_ = get_dataset_config_names(snake_case__ ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' , [ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] , ) def a ( snake_case__: List[Any] , snake_case__: Optional[int] , snake_case__: Any ): '''simple docstring''' lowercase_ = get_dataset_infos(snake_case__ ) assert list(infos.keys() ) == expected_configs lowercase_ = expected_configs[0] assert expected_config in infos lowercase_ = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def a ( snake_case__: str , snake_case__: int , snake_case__: Tuple ): '''simple docstring''' lowercase_ = get_dataset_infos(snake_case__ ) assert expected_config in infos lowercase_ = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def a ( snake_case__: List[str] , snake_case__: str , snake_case__: Optional[Any] ): '''simple docstring''' with pytest.raises(snake_case__ ): get_dataset_split_names(snake_case__ , config_name=snake_case__ )
30
def a ( snake_case__: list ): '''simple docstring''' if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] lowercase_ = [] def generate(snake_case__: int , snake_case__: list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even lowercase_ , lowercase_ = arr[k - 1], arr[i] else: # k is odd lowercase_ , lowercase_ = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": __a = input('Enter numbers separated by a comma:\n').strip() __a = [int(item) for item in user_input.split(',')] print(heaps(arr))
30
1
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : List[str] = logging.get_logger(__name__) A : List[Any] = { 'nvidia/segformer-b0-finetuned-ade-512-512': ( 'https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json' ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A ( UpperCAmelCase__ ): '''simple docstring''' A__ = '''segformer''' def __init__(self : Dict , _UpperCAmelCase : Optional[int]=3 , _UpperCAmelCase : Dict=4 , _UpperCAmelCase : Union[str, Any]=[2, 2, 2, 2] , _UpperCAmelCase : List[str]=[8, 4, 2, 1] , _UpperCAmelCase : str=[32, 64, 160, 256] , _UpperCAmelCase : Optional[int]=[7, 3, 3, 3] , _UpperCAmelCase : int=[4, 2, 2, 2] , _UpperCAmelCase : str=[1, 2, 5, 8] , _UpperCAmelCase : Optional[int]=[4, 4, 4, 4] , _UpperCAmelCase : List[Any]="gelu" , _UpperCAmelCase : Optional[Any]=0.0 , _UpperCAmelCase : Dict=0.0 , _UpperCAmelCase : Dict=0.1 , _UpperCAmelCase : List[str]=0.02 , _UpperCAmelCase : str=0.1 , _UpperCAmelCase : Optional[Any]=1E-6 , _UpperCAmelCase : Optional[Any]=256 , _UpperCAmelCase : Any=255 , **_UpperCAmelCase : str , ) -> Tuple: """simple docstring""" super().__init__(**_UpperCAmelCase ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( """Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be""" """ removed, as the behaviour will default to that of reshape_last_stage = True.""" , _UpperCAmelCase , ) lowercase__ = num_channels lowercase__ = num_encoder_blocks lowercase__ = depths lowercase__ = sr_ratios lowercase__ = hidden_sizes lowercase__ = patch_sizes lowercase__ = strides lowercase__ = mlp_ratios lowercase__ = num_attention_heads lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = classifier_dropout_prob lowercase__ = initializer_range lowercase__ = drop_path_rate lowercase__ = layer_norm_eps lowercase__ = decoder_hidden_size lowercase__ = kwargs.get("""reshape_last_stage""" , _UpperCAmelCase ) lowercase__ = semantic_loss_ignore_index class A ( UpperCAmelCase__ ): '''simple docstring''' A__ = version.parse('''1.11''' ) @property def lowerCamelCase__ (self : Optional[int] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def lowerCamelCase__ (self : Optional[int] ) -> float: """simple docstring""" return 1E-4 @property def lowerCamelCase__ (self : Optional[Any] ) -> int: """simple docstring""" return 12
361
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class A ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase__ (self : Any ) -> Union[str, Any]: """simple docstring""" lowercase__ = TFXLMRobertaModel.from_pretrained("""jplu/tf-xlm-roberta-base""" ) lowercase__ = { """input_ids""": tf.convert_to_tensor([[0, 2646, 1_0269, 83, 9_9942, 2]] , dtype=tf.intaa ), # "My dog is cute" """attention_mask""": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } lowercase__ = model(_UpperCAmelCase )["""last_hidden_state"""] lowercase__ = tf.TensorShape((1, 6, 768) ) self.assertEqual(output.shape , _UpperCAmelCase ) # compare the actual values for a slice. lowercase__ = tf.convert_to_tensor( [ [ [0.0_681_762, 0.10_894_451, 0.06_772_504], [-0.06_423_668, 0.02_366_615, 0.04_329_344], [-0.06_057_295, 0.09_974_135, -0.00_070_584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
146
0
"""simple docstring""" import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin _lowercase : Tuple = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right _lowercase : List[str] = 25_00_04 _lowercase : int = 25_00_20 @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( _lowerCAmelCase , unittest.TestCase ): a__ : Union[str, Any] = MBartaaTokenizer a__ : List[str] = MBartaaTokenizerFast a__ : Any = True a__ : List[str] = True def a ( self : str ): super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase = MBartaaTokenizer(_lowercase , src_lang='''en_XX''' , tgt_lang='''ro_RO''' , keep_accents=_lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def a ( self : Dict ): __UpperCAmelCase = '''<s>''' __UpperCAmelCase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowercase ) , _lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowercase ) , _lowercase ) def a ( self : Optional[Any] ): __UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(_lowercase ) , 10_54 ) def a ( self : Tuple ): self.assertEqual(self.get_tokenizer().vocab_size , 10_54 ) def a ( self : str ): __UpperCAmelCase = MBartaaTokenizer(_lowercase , src_lang='''en_XX''' , tgt_lang='''ro_RO''' , keep_accents=_lowercase ) __UpperCAmelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowercase ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) __UpperCAmelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _lowercase , [SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.'''] , ) __UpperCAmelCase = tokenizer.convert_tokens_to_ids(_lowercase ) self.assertListEqual( _lowercase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __UpperCAmelCase = tokenizer.convert_ids_to_tokens(_lowercase ) self.assertListEqual( _lowercase , [SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.'''] , ) @slow def a ( self : str ): # fmt: off __UpperCAmelCase = {'''input_ids''': [[25_00_04, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [25_00_04, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [25_00_04, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowercase , model_name='''facebook/mbart-large-50''' , revision='''d3913889c59cd5c9e456b269c376325eabad57e2''' , ) def a ( self : str ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __UpperCAmelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-mbart50''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = self.tokenizer_class.from_pretrained(_lowercase , **_lowercase ) __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = tokenizer_r.save_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.save_pretrained(_lowercase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) __UpperCAmelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(_lowercase , _lowercase ) # Checks everything loads correctly in the same way __UpperCAmelCase = tokenizer_r.from_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase , _lowercase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(_lowercase ) # Save tokenizer rust, legacy_format=True __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = tokenizer_r.save_pretrained(_lowercase , legacy_format=_lowercase ) __UpperCAmelCase = tokenizer_p.save_pretrained(_lowercase ) # Checks it save with the same files self.assertSequenceEqual(_lowercase , _lowercase ) # Checks everything loads correctly in the same way __UpperCAmelCase = tokenizer_r.from_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase , _lowercase ) ) shutil.rmtree(_lowercase ) # Save tokenizer rust, legacy_format=False __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = tokenizer_r.save_pretrained(_lowercase , legacy_format=_lowercase ) __UpperCAmelCase = tokenizer_p.save_pretrained(_lowercase ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __UpperCAmelCase = tokenizer_r.from_pretrained(_lowercase ) __UpperCAmelCase = tokenizer_p.from_pretrained(_lowercase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(_lowercase , _lowercase ) ) shutil.rmtree(_lowercase ) @require_torch @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( unittest.TestCase ): a__ : str = "facebook/mbart-large-50-one-to-many-mmt" a__ : Union[str, Any] = [ " UN Chief Says There Is No Military Solution in Syria", " Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.", ] a__ : Any = [ "Şeful ONU declară că nu există o soluţie militară în Siria", "Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei" " pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor" " face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.", ] a__ : Any = [EN_CODE, 8_274, 127_873, 25_916, 7, 8_622, 2_071, 438, 67_485, 53, 187_895, 23, 51_712, 2] @classmethod def a ( cls : Tuple ): __UpperCAmelCase = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''en_XX''' , tgt_lang='''ro_RO''' ) __UpperCAmelCase = 1 return cls def a ( self : Union[str, Any] ): self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ar_AR'''] , 25_00_01 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''en_EN'''] , 25_00_04 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ro_RO'''] , 25_00_20 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''mr_IN'''] , 25_00_38 ) def a ( self : Union[str, Any] ): __UpperCAmelCase = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _lowercase ) def a ( self : Optional[Any] ): self.assertIn(_lowercase , self.tokenizer.all_special_ids ) __UpperCAmelCase = [RO_CODE, 8_84, 90_19, 96, 9, 9_16, 8_67_92, 36, 1_87_43, 1_55_96, 5, 2] __UpperCAmelCase = self.tokenizer.decode(_lowercase , skip_special_tokens=_lowercase ) __UpperCAmelCase = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_lowercase ) self.assertEqual(_lowercase , _lowercase ) self.assertNotIn(self.tokenizer.eos_token , _lowercase ) def a ( self : Optional[Any] ): __UpperCAmelCase = ['''this is gunna be a long sentence ''' * 20] assert isinstance(src_text[0] , _lowercase ) __UpperCAmelCase = 10 __UpperCAmelCase = self.tokenizer(_lowercase , max_length=_lowercase , truncation=_lowercase ).input_ids[0] self.assertEqual(ids[0] , _lowercase ) self.assertEqual(ids[-1] , 2 ) self.assertEqual(len(_lowercase ) , _lowercase ) def a ( self : Optional[int] ): self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [25_00_53, 25_00_01] ) def a ( self : Union[str, Any] ): __UpperCAmelCase = tempfile.mkdtemp() __UpperCAmelCase = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(_lowercase ) __UpperCAmelCase = MBartaaTokenizer.from_pretrained(_lowercase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , _lowercase ) @require_torch def a ( self : Dict ): __UpperCAmelCase = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_lowercase , return_tensors='''pt''' ) __UpperCAmelCase = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def a ( self : Union[str, Any] ): __UpperCAmelCase = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=_lowercase , truncation=_lowercase , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) __UpperCAmelCase = shift_tokens_right(batch['''labels'''] , self.tokenizer.pad_token_id ) self.assertIsInstance(_lowercase , _lowercase ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) __UpperCAmelCase = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , _lowercase ) self.assertEqual(2 , batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def a ( self : Union[str, Any] ): __UpperCAmelCase = self.tokenizer(self.src_text , padding=_lowercase , truncation=_lowercase , max_length=3 , return_tensors='''pt''' ) __UpperCAmelCase = self.tokenizer( text_target=self.tgt_text , padding=_lowercase , truncation=_lowercase , max_length=10 , return_tensors='''pt''' ) __UpperCAmelCase = targets['''input_ids'''] __UpperCAmelCase = shift_tokens_right(_lowercase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def a ( self : Dict ): __UpperCAmelCase = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''en_XX''' , tgt_lang='''ar_AR''' ) self.assertEqual( nested_simplify(_lowercase ) , { # en_XX, A, test, EOS '''input_ids''': [[25_00_04, 62, 30_34, 2]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 25_00_01, } , )
332
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class _UpperCAmelCase ( unittest.TestCase ): def a ( self : Dict , _lowercase : Union[str, Any] ): for model_result in results.values(): for batch_size, sequence_length in zip(model_result['''bs'''] , model_result['''ss'''] ): __UpperCAmelCase = model_result['''result'''][batch_size][sequence_length] self.assertIsNotNone(_lowercase ) def a ( self : str ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def a ( self : List[str] ): __UpperCAmelCase = '''sgugger/tiny-distilbert-classification''' __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , only_pretrain_model=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def a ( self : str ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , torchscript=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Cant do half precision''' ) def a ( self : Optional[Any] ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , fpaa=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def a ( self : int ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = AutoConfig.from_pretrained(_lowercase ) # set architectures equal to `None` __UpperCAmelCase = None __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase , configs=[config] ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def a ( self : Tuple ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == '''cpu''' , '''Can\'t do half precision''' ) def a ( self : Optional[Any] ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , fpaa=_lowercase , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def a ( self : Any ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = AutoConfig.from_pretrained(_lowercase ) __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase , configs=[config] ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def a ( self : str ): __UpperCAmelCase = '''sshleifer/tinier_bart''' __UpperCAmelCase = AutoConfig.from_pretrained(_lowercase ) __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase , configs=[config] ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def a ( self : Union[str, Any] ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' __UpperCAmelCase = AutoConfig.from_pretrained(_lowercase ) __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase , configs=[config] ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def a ( self : int ): __UpperCAmelCase = '''sshleifer/tinier_bart''' __UpperCAmelCase = AutoConfig.from_pretrained(_lowercase ) __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase , configs=[config] ) __UpperCAmelCase = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def a ( self : Optional[Any] ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' with tempfile.TemporaryDirectory() as tmp_dir: __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , save_to_csv=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(_lowercase , '''inf_time.csv''' ) , train_memory_csv_file=os.path.join(_lowercase , '''train_mem.csv''' ) , inference_memory_csv_file=os.path.join(_lowercase , '''inf_mem.csv''' ) , train_time_csv_file=os.path.join(_lowercase , '''train_time.csv''' ) , env_info_csv_file=os.path.join(_lowercase , '''env.csv''' ) , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) benchmark.run() self.assertTrue(Path(os.path.join(_lowercase , '''inf_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowercase , '''train_time.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowercase , '''inf_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowercase , '''train_mem.csv''' ) ).exists() ) self.assertTrue(Path(os.path.join(_lowercase , '''env.csv''' ) ).exists() ) def a ( self : List[Any] ): __UpperCAmelCase = '''sshleifer/tiny-gpt2''' def _check_summary_is_not_empty(_lowercase : str ): self.assertTrue(hasattr(_lowercase , '''sequential''' ) ) self.assertTrue(hasattr(_lowercase , '''cumulative''' ) ) self.assertTrue(hasattr(_lowercase , '''current''' ) ) self.assertTrue(hasattr(_lowercase , '''total''' ) ) with tempfile.TemporaryDirectory() as tmp_dir: __UpperCAmelCase = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=_lowercase , inference=_lowercase , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(_lowercase , '''log.txt''' ) , log_print=_lowercase , trace_memory_line_by_line=_lowercase , multi_process=_lowercase , ) __UpperCAmelCase = PyTorchBenchmark(_lowercase ) __UpperCAmelCase = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(_lowercase , '''log.txt''' ) ).exists() )
332
1
'''simple docstring''' from __future__ import annotations _SCREAMING_SNAKE_CASE : int = "Muhammad Umer Farooq" _SCREAMING_SNAKE_CASE : str = "MIT" _SCREAMING_SNAKE_CASE : Any = "1.0.0" _SCREAMING_SNAKE_CASE : str = "Muhammad Umer Farooq" _SCREAMING_SNAKE_CASE : Union[str, Any] = "contact@muhammadumerfarooq.me" _SCREAMING_SNAKE_CASE : str = "Alpha" import re from html.parser import HTMLParser from urllib import parse import requests class _snake_case ( lowercase_ ): def __init__( self , a__ ) -> None: '''simple docstring''' super().__init__() snake_case_ = [] snake_case_ = domain def lowerCAmelCase__ ( self , a__ , a__ ) -> None: '''simple docstring''' if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: snake_case_ = parse.urljoin(self.domain , a__ ) self.urls.append(a__ ) def UpperCamelCase_( snake_case : str ): '''simple docstring''' return ".".join(get_sub_domain_name(snake_case ).split("." )[-2:] ) def UpperCamelCase_( snake_case : str ): '''simple docstring''' return parse.urlparse(snake_case ).netloc def UpperCamelCase_( snake_case : str = "https://github.com" ): '''simple docstring''' snake_case_ = get_domain_name(snake_case ) # Initialize the parser snake_case_ = Parser(snake_case ) try: # Open URL snake_case_ = requests.get(snake_case ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through snake_case_ = set() for link in parser.urls: # open URL. # read = requests.get(link) try: snake_case_ = requests.get(snake_case ) # Get the valid email. snake_case_ = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(snake_case ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(snake_case ) if __name__ == "__main__": _SCREAMING_SNAKE_CASE : List[str] = emails_from_url("https://github.com") print(F"{len(emails)} emails found:") print("\n".join(sorted(emails)))
92
'''simple docstring''' def UpperCamelCase_( snake_case : int = 1_0_0_0_0_0_0 ): '''simple docstring''' snake_case_ = set(range(3 , snake_case , 2 ) ) primes.add(2 ) for p in range(3 , snake_case , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , snake_case , snake_case ) ) ) snake_case_ = [float(snake_case ) for n in range(limit + 1 )] for p in primes: for n in range(snake_case , limit + 1 , snake_case ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F"{solution() = }")
92
1
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset from utils import logger class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = params UpperCAmelCase_ : int = np.array(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = np.array([len(lowercase_ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self , lowercase_ ): """simple docstring""" return (self.token_ids[index], self.lengths[index]) def __len__( self ): """simple docstring""" return len(self.lengths ) def UpperCamelCase__ ( self ): """simple docstring""" assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = self.params.max_model_input_size UpperCAmelCase_ : str = self.lengths > max_len logger.info(F"""Splitting {sum(lowercase_ )} too long sequences.""" ) def divide_chunks(lowercase_ , lowercase_ ): return [l[i : i + n] for i in range(0 , len(lowercase_ ) , lowercase_ )] UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : List[Any] = [] if self.params.mlm: UpperCAmelCase_ , UpperCAmelCase_ : str = self.params.special_tok_ids["cls_token"], self.params.special_tok_ids["sep_token"] else: UpperCAmelCase_ , UpperCAmelCase_ : int = self.params.special_tok_ids["bos_token"], self.params.special_tok_ids["eos_token"] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: UpperCAmelCase_ : Optional[int] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: UpperCAmelCase_ : str = np.insert(lowercase_ , 0 , lowercase_ ) if sub_s[-1] != sep_id: UpperCAmelCase_ : List[str] = np.insert(lowercase_ , len(lowercase_ ) , lowercase_ ) assert len(lowercase_ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(lowercase_ ) new_tok_ids.extend(lowercase_ ) new_lengths.extend([len(lowercase_ ) for l in sub_seqs] ) UpperCAmelCase_ : List[str] = np.array(lowercase_ ) UpperCAmelCase_ : List[str] = np.array(lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = len(self ) UpperCAmelCase_ : Tuple = self.lengths > 11 UpperCAmelCase_ : Any = self.token_ids[indices] UpperCAmelCase_ : List[Any] = self.lengths[indices] UpperCAmelCase_ : List[Any] = len(self ) logger.info(F"""Remove {init_size - new_size} too short (<=11 tokens) sequences.""" ) def UpperCamelCase__ ( self ): """simple docstring""" if "unk_token" not in self.params.special_tok_ids: return else: UpperCAmelCase_ : List[str] = self.params.special_tok_ids["unk_token"] UpperCAmelCase_ : Optional[Any] = len(self ) UpperCAmelCase_ : List[str] = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) UpperCAmelCase_ : List[Any] = (unk_occs / self.lengths) < 0.5 UpperCAmelCase_ : Tuple = self.token_ids[indices] UpperCAmelCase_ : str = self.lengths[indices] UpperCAmelCase_ : int = len(self ) logger.info(F"""Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).""" ) def UpperCamelCase__ ( self ): """simple docstring""" if not self.params.is_master: return logger.info(F"""{len(self )} sequences""" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [t[0] for t in batch] UpperCAmelCase_ : Tuple = [t[1] for t in batch] assert len(lowercase_ ) == len(lowercase_ ) # Max for paddings UpperCAmelCase_ : int = max(lowercase_ ) # Pad token ids if self.params.mlm: UpperCAmelCase_ : Optional[Any] = self.params.special_tok_ids["pad_token"] else: UpperCAmelCase_ : int = self.params.special_tok_ids["unk_token"] UpperCAmelCase_ : int = [list(t.astype(lowercase_ ) ) + [pad_idx] * (max_seq_len_ - len(lowercase_ )) for t in token_ids] assert len(tk_ ) == len(lowercase_ ) assert all(len(lowercase_ ) == max_seq_len_ for t in tk_ ) UpperCAmelCase_ : List[Any] = torch.tensor(tk_ ) # (bs, max_seq_len_) UpperCAmelCase_ : str = torch.tensor(lowercase_ ) # (bs) return tk_t, lg_t
61
"""simple docstring""" import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : str = checkpoint UpperCAmelCase_ : int = {} UpperCAmelCase_ : Optional[int] = vae_state_dict["encoder.conv_in.weight"] UpperCAmelCase_ : List[str] = vae_state_dict["encoder.conv_in.bias"] UpperCAmelCase_ : Optional[Any] = vae_state_dict["encoder.conv_out.weight"] UpperCAmelCase_ : Optional[int] = vae_state_dict["encoder.conv_out.bias"] UpperCAmelCase_ : List[Any] = vae_state_dict["encoder.norm_out.weight"] UpperCAmelCase_ : Union[str, Any] = vae_state_dict["encoder.norm_out.bias"] UpperCAmelCase_ : Any = vae_state_dict["decoder.conv_in.weight"] UpperCAmelCase_ : int = vae_state_dict["decoder.conv_in.bias"] UpperCAmelCase_ : Any = vae_state_dict["decoder.conv_out.weight"] UpperCAmelCase_ : Tuple = vae_state_dict["decoder.conv_out.bias"] UpperCAmelCase_ : List[Any] = vae_state_dict["decoder.norm_out.weight"] UpperCAmelCase_ : Tuple = vae_state_dict["decoder.norm_out.bias"] UpperCAmelCase_ : str = vae_state_dict["quant_conv.weight"] UpperCAmelCase_ : Optional[Any] = vae_state_dict["quant_conv.bias"] UpperCAmelCase_ : List[str] = vae_state_dict["post_quant_conv.weight"] UpperCAmelCase_ : List[Any] = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only UpperCAmelCase_ : Optional[Any] = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) UpperCAmelCase_ : Optional[Any] = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase_ : Dict = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) UpperCAmelCase_ : Optional[int] = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } for i in range(__lowerCamelCase ): UpperCAmelCase_ : Any = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ : Dict = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase_ : Dict = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase_ : List[str] = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : List[Any] = [key for key in vae_state_dict if "encoder.mid.block" in key] UpperCAmelCase_ : Tuple = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase_ : str = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase_ : List[Any] = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = [key for key in vae_state_dict if "encoder.mid.attn" in key] UpperCAmelCase_ : Union[str, Any] = renew_vae_attention_paths(__lowerCamelCase ) UpperCAmelCase_ : int = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) for i in range(__lowerCamelCase ): UpperCAmelCase_ : Optional[Any] = num_up_blocks - 1 - i UpperCAmelCase_ : Any = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ : str = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase_ : Optional[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase_ : Dict = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : List[str] = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = [key for key in vae_state_dict if "decoder.mid.block" in key] UpperCAmelCase_ : List[Any] = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase_ : str = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase_ : Tuple = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Tuple = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Optional[int] = [key for key in vae_state_dict if "decoder.mid.attn" in key] UpperCAmelCase_ : Any = renew_vae_attention_paths(__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) return new_checkpoint def __a ( __lowerCamelCase, __lowerCamelCase, ): # Only support V1 UpperCAmelCase_ : List[str] = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) UpperCAmelCase_ : List[Any] = io.BytesIO(r.content ) UpperCAmelCase_ : Any = OmegaConf.load(__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = 512 UpperCAmelCase_ : Optional[Any] = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open UpperCAmelCase_ : int = {} with safe_open(__lowerCamelCase, framework="pt", device="cpu" ) as f: for key in f.keys(): UpperCAmelCase_ : Tuple = f.get_tensor(__lowerCamelCase ) else: UpperCAmelCase_ : Any = torch.load(__lowerCamelCase, map_location=__lowerCamelCase )["state_dict"] # Convert the VAE model. UpperCAmelCase_ : Dict = create_vae_diffusers_config(__lowerCamelCase, image_size=__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = custom_convert_ldm_vae_checkpoint(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : int = AutoencoderKL(**__lowerCamelCase ) vae.load_state_dict(__lowerCamelCase ) vae.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument('--vae_pt_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') _a = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
61
1
"""simple docstring""" # flake8: noqa # Lint as: python3 from typing import Dict, List, Optional, Type from .. import config from ..utils import logging from .formatting import ( ArrowFormatter, CustomFormatter, Formatter, PandasFormatter, PythonFormatter, TensorFormatter, format_table, query_table, ) from .np_formatter import NumpyFormatter lowerCamelCase_ : str = logging.get_logger(__name__) lowerCamelCase_ : Dict[Optional[str], Type[Formatter]] = {} lowerCamelCase_ : Dict[Optional[str], str] = {} lowerCamelCase_ : Dict[Optional[str], Exception] = {} def _A ( lowercase , lowercase , lowercase = None , ): """simple docstring""" a =aliases if aliases is not None else [] if format_type in _FORMAT_TYPES: logger.warning( f'''Overwriting format type \'{format_type}\' ({_FORMAT_TYPES[format_type].__name__} -> {formatter_cls.__name__})''' ) a =formatter_cls for alias in set(aliases + [format_type] ): if alias in _FORMAT_TYPES_ALIASES: logger.warning( f'''Overwriting format type alias \'{alias}\' ({_FORMAT_TYPES_ALIASES[alias]} -> {format_type})''' ) a =format_type def _A ( lowercase , lowercase , lowercase = None ): """simple docstring""" a =aliases if aliases is not None else [] for alias in set(aliases + [format_type] ): a =unavailable_error # Here we define all the available formatting functions that can be used by `Dataset.set_format` _register_formatter(PythonFormatter, None, aliases=["""python"""]) _register_formatter(ArrowFormatter, """arrow""", aliases=["""pa""", """pyarrow"""]) _register_formatter(NumpyFormatter, """numpy""", aliases=["""np"""]) _register_formatter(PandasFormatter, """pandas""", aliases=["""pd"""]) _register_formatter(CustomFormatter, """custom""") if config.TORCH_AVAILABLE: from .torch_formatter import TorchFormatter _register_formatter(TorchFormatter, """torch""", aliases=["""pt""", """pytorch"""]) else: lowerCamelCase_ : Dict = ValueError("""PyTorch needs to be installed to be able to return PyTorch tensors.""") _register_unavailable_formatter(_torch_error, """torch""", aliases=["""pt""", """pytorch"""]) if config.TF_AVAILABLE: from .tf_formatter import TFFormatter _register_formatter(TFFormatter, """tensorflow""", aliases=["""tf"""]) else: lowerCamelCase_ : Tuple = ValueError("""Tensorflow needs to be installed to be able to return Tensorflow tensors.""") _register_unavailable_formatter(_tf_error, """tensorflow""", aliases=["""tf"""]) if config.JAX_AVAILABLE: from .jax_formatter import JaxFormatter _register_formatter(JaxFormatter, """jax""", aliases=[]) else: lowerCamelCase_ : Optional[int] = ValueError("""JAX needs to be installed to be able to return JAX arrays.""") _register_unavailable_formatter(_jax_error, """jax""", aliases=[]) def _A ( lowercase ): """simple docstring""" if format_type in _FORMAT_TYPES_ALIASES: return _FORMAT_TYPES_ALIASES[format_type] else: return format_type def _A ( lowercase , **lowercase ): """simple docstring""" a =get_format_type_from_alias(lowercase ) if format_type in _FORMAT_TYPES: return _FORMAT_TYPES[format_type](**lowercase ) if format_type in _FORMAT_TYPES_ALIASES_UNAVAILABLE: raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type] else: raise ValueError( f'''Return type should be None or selected in {list(type for type in _FORMAT_TYPES.keys() if type != None )}, but got \'{format_type}\'''' )
215
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCamelCase_ : List[Any] = logging.get_logger(__name__) if is_vision_available(): import PIL class __A ( _SCREAMING_SNAKE_CASE ): """simple docstring""" __lowerCAmelCase = ["pixel_values"] def __init__( self , __A = True , __A = None , __A = PILImageResampling.BICUBIC , __A = True , __A = None , __A = True , __A = 1 / 255 , __A = True , __A = None , __A = None , __A = True , **__A , ) -> None: super().__init__(**__A ) a =size if size is not None else {'''shortest_edge''': 224} a =get_size_dict(__A , default_to_square=__A ) a =crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} a =get_size_dict(__A , default_to_square=__A , param_name='''crop_size''' ) a =do_resize a =size a =resample a =do_center_crop a =crop_size a =do_rescale a =rescale_factor a =do_normalize a =image_mean if image_mean is not None else OPENAI_CLIP_MEAN a =image_std if image_std is not None else OPENAI_CLIP_STD a =do_convert_rgb def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = PILImageResampling.BICUBIC , __A = None , **__A , ) -> np.ndarray: a =get_size_dict(__A , default_to_square=__A ) if "shortest_edge" not in size: raise ValueError(f'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''' ) a =get_resize_output_image_size(__A , size=size['''shortest_edge'''] , default_to_square=__A ) return resize(__A , size=__A , resample=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = None , **__A , ) -> np.ndarray: a =get_size_dict(__A ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` parameter must contain the keys (height, width). Got {size.keys()}''' ) return center_crop(__A , size=(size['''height'''], size['''width''']) , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A = None , **__A , ) -> Any: return rescale(__A , scale=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A , __A , __A = None , **__A , ) -> np.ndarray: return normalize(__A , mean=__A , std=__A , data_format=__A , **__A ) def SCREAMING_SNAKE_CASE ( self , __A , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = None , __A = ChannelDimension.FIRST , **__A , ) -> PIL.Image.Image: a =do_resize if do_resize is not None else self.do_resize a =size if size is not None else self.size a =get_size_dict(__A , param_name='''size''' , default_to_square=__A ) a =resample if resample is not None else self.resample a =do_center_crop if do_center_crop is not None else self.do_center_crop a =crop_size if crop_size is not None else self.crop_size a =get_size_dict(__A , param_name='''crop_size''' , default_to_square=__A ) a =do_rescale if do_rescale is not None else self.do_rescale a =rescale_factor if rescale_factor is not None else self.rescale_factor a =do_normalize if do_normalize is not None else self.do_normalize a =image_mean if image_mean is not None else self.image_mean a =image_std if image_std is not None else self.image_std a =do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb a =make_list_of_images(__A ) if not valid_images(__A ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: a =[convert_to_rgb(__A ) for image in images] # All transformations expect numpy arrays. a =[to_numpy_array(__A ) for image in images] if do_resize: a =[self.resize(image=__A , size=__A , resample=__A ) for image in images] if do_center_crop: a =[self.center_crop(image=__A , size=__A ) for image in images] if do_rescale: a =[self.rescale(image=__A , scale=__A ) for image in images] if do_normalize: a =[self.normalize(image=__A , mean=__A , std=__A ) for image in images] a =[to_channel_dimension_format(__A , __A ) for image in images] a ={'''pixel_values''': images} return BatchFeature(data=__A , tensor_type=__A )
215
1
import argparse import os import re import packaging.version __lowerCamelCase : str = '''examples/''' __lowerCamelCase : Dict = { '''examples''': (re.compile(R'''^check_min_version\("[^"]+"\)\s*$''', re.MULTILINE), '''check_min_version("VERSION")\n'''), '''init''': (re.compile(R'''^__version__\s+=\s+"([^"]+)"\s*$''', re.MULTILINE), '''__version__ = "VERSION"\n'''), '''setup''': (re.compile(R'''^(\s*)version\s*=\s*"[^"]+",''', re.MULTILINE), R'''\1version="VERSION",'''), '''doc''': (re.compile(R'''^(\s*)release\s*=\s*"[^"]+"$''', re.MULTILINE), '''release = "VERSION"\n'''), } __lowerCamelCase : Tuple = { '''init''': '''src/diffusers/__init__.py''', '''setup''': '''setup.py''', } __lowerCamelCase : str = '''README.md''' def _snake_case ( lowerCAmelCase : str , lowerCAmelCase : str , lowerCAmelCase : Union[str, Any] ): """simple docstring""" with open(lowerCAmelCase , "r" , encoding="utf-8" , newline="\n" ) as f: SCREAMING_SNAKE_CASE_ : List[Any] = f.read() SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str = REPLACE_PATTERNS[pattern] SCREAMING_SNAKE_CASE_ : str = replace.replace("VERSION" , lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[int] = re_pattern.sub(lowerCAmelCase , lowerCAmelCase ) with open(lowerCAmelCase , "w" , encoding="utf-8" , newline="\n" ) as f: f.write(lowerCAmelCase ) def _snake_case ( lowerCAmelCase : List[str] ): """simple docstring""" for folder, directories, fnames in os.walk(lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("research_projects" ) if "legacy" in directories: directories.remove("legacy" ) for fname in fnames: if fname.endswith(".py" ): update_version_in_file(os.path.join(lowerCAmelCase , lowerCAmelCase ) , lowerCAmelCase , pattern="examples" ) def _snake_case ( lowerCAmelCase : str , lowerCAmelCase : Optional[int]=False ): """simple docstring""" for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) if not patch: update_version_in_examples(lowerCAmelCase ) def _snake_case ( ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = "🤗 Transformers currently provides the following architectures" SCREAMING_SNAKE_CASE_ : Tuple = "1. Want to contribute a new model?" with open(lowerCAmelCase , "r" , encoding="utf-8" , newline="\n" ) as f: SCREAMING_SNAKE_CASE_ : Optional[Any] = f.readlines() # Find the start of the list. SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 SCREAMING_SNAKE_CASE_ : Any = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("1." ): SCREAMING_SNAKE_CASE_ : Any = lines[index].replace( "https://huggingface.co/docs/diffusers/main/model_doc" , "https://huggingface.co/docs/diffusers/model_doc" , ) index += 1 with open(lowerCAmelCase , "w" , encoding="utf-8" , newline="\n" ) as f: f.writelines(lowerCAmelCase ) def _snake_case ( ): """simple docstring""" with open(REPLACE_FILES["init"] , "r" ) as f: SCREAMING_SNAKE_CASE_ : Any = f.read() SCREAMING_SNAKE_CASE_ : Optional[int] = REPLACE_PATTERNS["init"][0].search(lowerCAmelCase ).groups()[0] return packaging.version.parse(lowerCAmelCase ) def _snake_case ( lowerCAmelCase : int=False ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = get_version() if patch and default_version.is_devrelease: raise ValueError("Can't create a patch version from the dev branch, checkout a released version!" ) if default_version.is_devrelease: SCREAMING_SNAKE_CASE_ : str = default_version.base_version elif patch: SCREAMING_SNAKE_CASE_ : Tuple = f'{default_version.major}.{default_version.minor}.{default_version.micro + 1}' else: SCREAMING_SNAKE_CASE_ : int = f'{default_version.major}.{default_version.minor + 1}.0' # Now let's ask nicely if that's the right one. SCREAMING_SNAKE_CASE_ : List[str] = input(f'Which version are you releasing? [{default_version}]' ) if len(lowerCAmelCase ) == 0: SCREAMING_SNAKE_CASE_ : Optional[int] = default_version print(f'Updating version to {version}.' ) global_version_update(lowerCAmelCase , patch=lowerCAmelCase ) def _snake_case ( ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = get_version() SCREAMING_SNAKE_CASE_ : List[Any] = f'{current_version.major}.{current_version.minor + 1}.0.dev0' SCREAMING_SNAKE_CASE_ : int = current_version.base_version # Check with the user we got that right. SCREAMING_SNAKE_CASE_ : Dict = input(f'Which version are we developing now? [{dev_version}]' ) if len(lowerCAmelCase ) == 0: SCREAMING_SNAKE_CASE_ : List[Any] = dev_version print(f'Updating version to {version}.' ) global_version_update(lowerCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": __lowerCamelCase : Dict = argparse.ArgumentParser() parser.add_argument('''--post_release''', action='''store_true''', help='''Whether this is pre or post release.''') parser.add_argument('''--patch''', action='''store_true''', help='''Whether or not this is a patch release.''') __lowerCamelCase : Tuple = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('''Nothing to do after a patch :-)''') else: post_release_work()
18
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): __UpperCAmelCase = "pt" elif is_tf_available(): __UpperCAmelCase = "tf" else: __UpperCAmelCase = "jax" class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ =ByTaTokenizer UpperCAmelCase_ =False def _UpperCamelCase ( self ) -> Tuple: super().setUp() SCREAMING_SNAKE_CASE_ = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _UpperCamelCase ( self ) -> List[str]: return ByTaTokenizer.from_pretrained('''google/byt5-small''' ) def _UpperCamelCase ( self , **_A ) -> ByTaTokenizer: return self.tokenizer_class.from_pretrained(self.tmpdirname , **_A ) def _UpperCamelCase ( self , _A , _A=False , _A=20 , _A=5 ) -> Tuple[str, list]: # XXX The default common tokenizer tests assume that every ID is decodable on its own. # This assumption is invalid for ByT5 because single bytes might not be # valid utf-8 (byte 128 for instance). # Here we're overriding the smallest possible method to provide # a clean sequence without making the same assumption. SCREAMING_SNAKE_CASE_ = [] for i in range(len(_A ) ): try: SCREAMING_SNAKE_CASE_ = tokenizer.decode([i] , clean_up_tokenization_spaces=_A ) except UnicodeDecodeError: pass toks.append((i, tok) ) SCREAMING_SNAKE_CASE_ = list(filter(lambda _A : re.match(R'''^[ a-zA-Z]+$''' , t[1] ) , _A ) ) SCREAMING_SNAKE_CASE_ = list(filter(lambda _A : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=_A ) , _A ) ) if max_length is not None and len(_A ) > max_length: SCREAMING_SNAKE_CASE_ = toks[:max_length] if min_length is not None and len(_A ) < min_length and len(_A ) > 0: while len(_A ) < min_length: SCREAMING_SNAKE_CASE_ = toks + toks # toks_str = [t[1] for t in toks] SCREAMING_SNAKE_CASE_ = [t[0] for t in toks] # Ensure consistency SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A , clean_up_tokenization_spaces=_A ) if " " not in output_txt and len(_A ) > 1: SCREAMING_SNAKE_CASE_ = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=_A ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=_A ) ) if with_prefix_space: SCREAMING_SNAKE_CASE_ = ''' ''' + output_txt SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) return output_txt, output_ids def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = tokenizer(['''hi</s>''', '''I went to the gym</s>''', '''</s>'''] ) SCREAMING_SNAKE_CASE_ = tokenizer(['''hi''', '''I went to the gym''', ''''''] ) self.assertListEqual(batch_with_eos_added['''input_ids'''] , batch_without_eos_added['''input_ids'''] ) def _UpperCamelCase ( self ) -> Any: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = '''Unicode €.''' SCREAMING_SNAKE_CASE_ = tokenizer(_A ) SCREAMING_SNAKE_CASE_ = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A ) self.assertEqual(_A , '''Unicode €.</s>''' ) SCREAMING_SNAKE_CASE_ = tokenizer('''e è é ê ë''' ) SCREAMING_SNAKE_CASE_ = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded['''input_ids'''] , _A ) # decoding SCREAMING_SNAKE_CASE_ = tokenizer.decode(_A ) self.assertEqual(_A , '''e è é ê ë</s>''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''e è é ê ë</s>''' ) def _UpperCamelCase ( self ) -> List[str]: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off SCREAMING_SNAKE_CASE_ = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on SCREAMING_SNAKE_CASE_ = tokenizer(_A , padding=_A , return_tensors=_A ) self.assertIsInstance(_A , _A ) if FRAMEWORK != "jax": SCREAMING_SNAKE_CASE_ = list(batch.input_ids.numpy()[0] ) else: SCREAMING_SNAKE_CASE_ = list(batch.input_ids.tolist()[0] ) self.assertListEqual(_A , _A ) self.assertEqual((2, 37) , batch.input_ids.shape ) self.assertEqual((2, 37) , batch.attention_mask.shape ) def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] SCREAMING_SNAKE_CASE_ = tokenizer(_A , padding=_A , return_tensors=_A ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , _A ) self.assertIn('''attention_mask''' , _A ) self.assertNotIn('''decoder_input_ids''' , _A ) self.assertNotIn('''decoder_attention_mask''' , _A ) def _UpperCamelCase ( self ) -> Tuple: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = [ '''Summary of the text.''', '''Another summary.''', ] SCREAMING_SNAKE_CASE_ = tokenizer( text_target=_A , max_length=32 , padding='''max_length''' , truncation=_A , return_tensors=_A ) self.assertEqual(32 , targets['''input_ids'''].shape[1] ) def _UpperCamelCase ( self ) -> List[Any]: SCREAMING_SNAKE_CASE_ = self.ta_base_tokenizer SCREAMING_SNAKE_CASE_ = ['''A long paragraph for summarization. </s>'''] SCREAMING_SNAKE_CASE_ = ['''Summary of the text. </s>'''] # fmt: off SCREAMING_SNAKE_CASE_ = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] SCREAMING_SNAKE_CASE_ = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on SCREAMING_SNAKE_CASE_ = tokenizer(_A , text_target=_A ) self.assertEqual(_A , batch['''input_ids'''][0] ) self.assertEqual(_A , batch['''labels'''][0] ) def _UpperCamelCase ( self ) -> Dict: # safety check on max_len default value so we are sure the test works SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = ''' He is very happy, UNwant\u00E9d,running''' SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A ) SCREAMING_SNAKE_CASE_ = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) shutil.rmtree(_A ) SCREAMING_SNAKE_CASE_ = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): # Isolate this from the other tests because we save additional tokens/etc SCREAMING_SNAKE_CASE_ = tempfile.mkdtemp() SCREAMING_SNAKE_CASE_ = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) SCREAMING_SNAKE_CASE_ = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) SCREAMING_SNAKE_CASE_ = tokenizer.encode(_A , add_special_tokens=_A ) tokenizer.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A ) SCREAMING_SNAKE_CASE_ = after_tokenizer.encode(_A , add_special_tokens=_A ) self.assertListEqual(_A , _A ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) SCREAMING_SNAKE_CASE_ = tokenizer.__class__.from_pretrained(_A , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(_A ) def _UpperCamelCase ( self ) -> int: SCREAMING_SNAKE_CASE_ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_A ) with open(os.path.join(_A , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: SCREAMING_SNAKE_CASE_ = json.load(_A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: SCREAMING_SNAKE_CASE_ = json.load(_A ) SCREAMING_SNAKE_CASE_ = [F'''<extra_id_{i}>''' for i in range(125 )] SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(_A , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) with open(os.path.join(_A , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(_A , _A ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained( _A , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained SCREAMING_SNAKE_CASE_ = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=_A )] SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained( _A , additional_special_tokens=_A , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_A ) SCREAMING_SNAKE_CASE_ = tokenizer_class.from_pretrained(_A ) self.assertTrue(tokenizer.decode([255] ) == '''''' ) def _UpperCamelCase ( self ) -> int: pass def _UpperCamelCase ( self ) -> Any: pass def _UpperCamelCase ( self ) -> Any: pass def _UpperCamelCase ( self ) -> Optional[int]: pass def _UpperCamelCase ( self ) -> Union[str, Any]: # The default common tokenizer tests uses invalid tokens for ByT5 that can only accept one-character strings # and special added tokens as tokens SCREAMING_SNAKE_CASE_ = self.get_tokenizers(fast=_A , do_lower_case=_A ) for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ = ['''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''x''', '''t''', '''</s>'''] SCREAMING_SNAKE_CASE_ = tokenizer.convert_tokens_to_string(_A ) self.assertIsInstance(_A , _A ) def _UpperCamelCase ( self ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'''{tokenizer.__class__.__name__}''' ): SCREAMING_SNAKE_CASE_ = [ '''bos_token''', '''eos_token''', '''unk_token''', '''sep_token''', '''pad_token''', '''cls_token''', '''mask_token''', ] SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = tokenizer.convert_ids_to_tokens( _A , skip_special_tokens=_A ) for attr in attributes_list: setattr(_A , attr + '''_id''' , _A ) self.assertEqual(getattr(_A , _A ) , _A ) self.assertEqual(getattr(_A , attr + '''_id''' ) , _A ) setattr(_A , attr + '''_id''' , _A ) self.assertEqual(getattr(_A , _A ) , _A ) self.assertEqual(getattr(_A , attr + '''_id''' ) , _A ) setattr(_A , '''additional_special_tokens_ids''' , [] ) self.assertListEqual(getattr(_A , '''additional_special_tokens''' ) , [] ) self.assertListEqual(getattr(_A , '''additional_special_tokens_ids''' ) , [] ) setattr(_A , '''additional_special_tokens_ids''' , [token_id_to_test_setters] ) self.assertListEqual(getattr(_A , '''additional_special_tokens''' ) , [token_to_test_setters] ) self.assertListEqual(getattr(_A , '''additional_special_tokens_ids''' ) , [token_id_to_test_setters] )
299
0
from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 __UpperCAmelCase : Dict = { # 1536-bit 5: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 2048-bit 14: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AACAA68FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 3072-bit 15: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 4096-bit 16: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7" + "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA" + "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6" + "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED" + "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9" + "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199" + "FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 6144-bit 17: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08" + "8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B" + "302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9" + "A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6" + "49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8" + "FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C" + "180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718" + "3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D" + "04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D" + "B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226" + "1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC" + "E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26" + "99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB" + "04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2" + "233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127" + "D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492" + "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406" + "AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918" + "DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151" + "2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03" + "F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F" + "BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA" + "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B" + "B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632" + "387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E" + "6DCC4024FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 8192-bit 18: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7" + "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA" + "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6" + "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED" + "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9" + "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492" + "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD" + "F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831" + "179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B" + "DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF" + "5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6" + "D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3" + "23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA" + "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328" + "06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C" + "DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE" + "12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4" + "38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300" + "741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568" + "3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9" + "22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B" + "4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A" + "062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36" + "4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1" + "B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92" + "4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47" + "9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71" + "60C980DD98EDD3DFFFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, } class __snake_case : '''simple docstring''' def __init__( self : Optional[int] , A : int = 14 ): if group not in primes: raise ValueError("""Unsupported Group""" ) __snake_case: Any = primes[group]["""prime"""] __snake_case: List[str] = primes[group]["""generator"""] __snake_case: str = int(hexlify(urandom(32 ) ) , base=16 ) def UpperCAmelCase__ ( self : List[str] ): return hex(self.__private_key )[2:] def UpperCAmelCase__ ( self : str ): __snake_case: Union[str, Any] = pow(self.generator , self.__private_key , self.prime ) return hex(A )[2:] def UpperCAmelCase__ ( self : Any , A : int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= key <= self.prime - 2 and pow(A , (self.prime - 1) // 2 , self.prime ) == 1 ) def UpperCAmelCase__ ( self : Dict , A : str ): __snake_case: Tuple = int(A , base=16 ) if not self.is_valid_public_key(A ): raise ValueError("""Invalid public key""" ) __snake_case: List[str] = pow(A , self.__private_key , self.prime ) return shaaaa(str(A ).encode() ).hexdigest() @staticmethod def UpperCAmelCase__ ( A : int , A : int ): # check if the other public key is valid based on NIST SP800-56 return ( 2 <= remote_public_key_str <= prime - 2 and pow(A , (prime - 1) // 2 , A ) == 1 ) @staticmethod def UpperCAmelCase__ ( A : str , A : str , A : int = 14 ): __snake_case: Optional[int] = int(A , base=16 ) __snake_case: Any = int(A , base=16 ) __snake_case: List[str] = primes[group]["""prime"""] if not DiffieHellman.is_valid_public_key_static(A , A ): raise ValueError("""Invalid public key""" ) __snake_case: List[str] = pow(A , A , A ) return shaaaa(str(A ).encode() ).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
355
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class __snake_case ( unittest.TestCase ): '''simple docstring''' @parameterized.expand([(None,), ("""foo.json""",)] ) def UpperCAmelCase__ ( self : List[str] , A : Optional[Any] ): __snake_case: Any = GenerationConfig( do_sample=A , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(A , config_name=A ) __snake_case: Optional[int] = GenerationConfig.from_pretrained(A , config_name=A ) # Checks parameters that were specified self.assertEqual(loaded_config.do_sample , A ) self.assertEqual(loaded_config.temperature , 0.7 ) self.assertEqual(loaded_config.length_penalty , 1.0 ) self.assertEqual(loaded_config.bad_words_ids , [[1, 2, 3], [4, 5]] ) # Checks parameters that were not specified (defaults) self.assertEqual(loaded_config.top_k , 50 ) self.assertEqual(loaded_config.max_length , 20 ) self.assertEqual(loaded_config.max_time , A ) def UpperCAmelCase__ ( self : Dict ): __snake_case: str = AutoConfig.from_pretrained("""gpt2""" ) __snake_case: Any = GenerationConfig.from_model_config(A ) __snake_case: str = GenerationConfig() # The generation config has loaded a few non-default parameters from the model config self.assertNotEqual(A , A ) # One of those parameters is eos_token_id -- check if it matches self.assertNotEqual(generation_config_from_model.eos_token_id , default_generation_config.eos_token_id ) self.assertEqual(generation_config_from_model.eos_token_id , model_config.eos_token_id ) def UpperCAmelCase__ ( self : str ): __snake_case: List[str] = GenerationConfig() __snake_case: Tuple = { """max_new_tokens""": 1_024, """foo""": """bar""", } __snake_case: List[str] = copy.deepcopy(A ) __snake_case: Optional[int] = generation_config.update(**A ) # update_kwargs was not modified (no side effects) self.assertEqual(A , A ) # update_kwargs was used to update the config on valid attributes self.assertEqual(generation_config.max_new_tokens , 1_024 ) # `.update()` returns a dictionary of unused kwargs self.assertEqual(A , {"""foo""": """bar"""} ) def UpperCAmelCase__ ( self : Tuple ): __snake_case: List[str] = GenerationConfig() __snake_case: Optional[int] = """bar""" with tempfile.TemporaryDirectory("""test-generation-config""" ) as tmp_dir: generation_config.save_pretrained(A ) __snake_case: Any = GenerationConfig.from_pretrained(A ) # update_kwargs was used to update the config on valid attributes self.assertEqual(new_config.foo , """bar""" ) __snake_case: int = GenerationConfig.from_model_config(A ) assert not hasattr(A , """foo""" ) # no new kwargs should be initialized if from config def UpperCAmelCase__ ( self : Dict ): __snake_case: Dict = GenerationConfig() self.assertEqual(default_config.temperature , 1.0 ) self.assertEqual(default_config.do_sample , A ) self.assertEqual(default_config.num_beams , 1 ) __snake_case: Union[str, Any] = GenerationConfig( do_sample=A , temperature=0.7 , length_penalty=1.0 , bad_words_ids=[[1, 2, 3], [4, 5]] , ) self.assertEqual(config.temperature , 0.7 ) self.assertEqual(config.do_sample , A ) self.assertEqual(config.num_beams , 1 ) with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(A ) __snake_case: Tuple = GenerationConfig.from_pretrained(A , temperature=1.0 ) self.assertEqual(loaded_config.temperature , 1.0 ) self.assertEqual(loaded_config.do_sample , A ) self.assertEqual(loaded_config.num_beams , 1 ) # default value @is_staging_test class __snake_case ( unittest.TestCase ): '''simple docstring''' @classmethod def UpperCAmelCase__ ( cls : List[str] ): __snake_case: Optional[int] = TOKEN HfFolder.save_token(A ) @classmethod def UpperCAmelCase__ ( cls : List[Any] ): try: delete_repo(token=cls._token , repo_id="""test-generation-config""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-generation-config-org""" ) except HTTPError: pass def UpperCAmelCase__ ( self : Tuple ): __snake_case: Optional[int] = GenerationConfig( do_sample=A , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("""test-generation-config""" , use_auth_token=self._token ) __snake_case: str = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A , getattr(A , A ) ) # Reset repo delete_repo(token=self._token , repo_id="""test-generation-config""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( A , repo_id="""test-generation-config""" , push_to_hub=A , use_auth_token=self._token ) __snake_case: Optional[Any] = GenerationConfig.from_pretrained(f'''{USER}/test-generation-config''' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A , getattr(A , A ) ) def UpperCAmelCase__ ( self : List[Any] ): __snake_case: Union[str, Any] = GenerationConfig( do_sample=A , temperature=0.7 , length_penalty=1.0 , ) config.push_to_hub("""valid_org/test-generation-config-org""" , use_auth_token=self._token ) __snake_case: int = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A , getattr(A , A ) ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-generation-config-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( A , repo_id="""valid_org/test-generation-config-org""" , push_to_hub=A , use_auth_token=self._token ) __snake_case: Optional[int] = GenerationConfig.from_pretrained("""valid_org/test-generation-config-org""" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(A , getattr(A , A ) )
293
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __a = { "configuration_altclip": [ "ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "AltCLIPConfig", "AltCLIPTextConfig", "AltCLIPVisionConfig", ], "processing_altclip": ["AltCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "AltCLIPPreTrainedModel", "AltCLIPModel", "AltCLIPTextModel", "AltCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
66
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __a = { "configuration_mask2former": [ "MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "Mask2FormerConfig", ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = ["Mask2FormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a = [ "MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "Mask2FormerForUniversalSegmentation", "Mask2FormerModel", "Mask2FormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_maskaformer import MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskaFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_maskaformer import MaskaFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskaformer import ( MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskaFormerForUniversalSegmentation, MaskaFormerModel, MaskaFormerPreTrainedModel, ) else: import sys __a = _LazyModule(__name__, globals()["__file__"], _import_structure)
66
1
"""simple docstring""" from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets __A = "\\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n" __A = "\\nGLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n" __A = "\nCompute GLUE evaluation metric associated to each GLUE dataset.\nArgs:\n predictions: list of predictions to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\nReturns: depending on the GLUE subset, one or several of:\n \"accuracy\": Accuracy\n \"f1\": F1 score\n \"pearson\": Pearson Correlation\n \"spearmanr\": Spearman Correlation\n \"matthews_correlation\": Matthew Correlation\nExamples:\n\n >>> glue_metric = datasets.load_metric('glue', 'sst2') # 'sst2' or any of [\"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"]\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'mrpc') # 'mrpc' or 'qqp'\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'accuracy': 1.0, 'f1': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'stsb')\n >>> references = [0., 1., 2., 3., 4., 5.]\n >>> predictions = [0., 1., 2., 3., 4., 5.]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print({\"pearson\": round(results[\"pearson\"], 2), \"spearmanr\": round(results[\"spearmanr\"], 2)})\n {'pearson': 1.0, 'spearmanr': 1.0}\n\n >>> glue_metric = datasets.load_metric('glue', 'cola')\n >>> references = [0, 1]\n >>> predictions = [0, 1]\n >>> results = glue_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'matthews_correlation': 1.0}\n" def a__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[Any]: return float((preds == labels).mean() ) def a__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> Optional[int]: __lowerCAmelCase: Optional[int] = simple_accuracy(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __lowerCAmelCase: List[Any] = float(fa_score(y_true=__SCREAMING_SNAKE_CASE , y_pred=__SCREAMING_SNAKE_CASE ) ) return { "accuracy": acc, "f1": fa, } def a__ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> List[str]: __lowerCAmelCase: Any = float(pearsonr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )[0] ) __lowerCAmelCase: List[str] = float(spearmanr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class snake_case ( datasets.Metric ): def lowercase_ ( self : Dict)-> Any: '''simple docstring''' if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]") return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "stsb" else "float32"), "references": datasets.Value("int64" if self.config_name != "stsb" else "float32"), }) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def lowercase_ ( self : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : List[Any])-> Optional[Any]: '''simple docstring''' if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(UpperCamelCase__ , UpperCamelCase__)} elif self.config_name == "stsb": return pearson_and_spearman(UpperCamelCase__ , UpperCamelCase__) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(UpperCamelCase__ , UpperCamelCase__) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(UpperCamelCase__ , UpperCamelCase__)} else: raise KeyError( "You should supply a configuration name selected in " "[\"sst2\", \"mnli\", \"mnli_mismatched\", \"mnli_matched\", " "\"cola\", \"stsb\", \"mrpc\", \"qqp\", \"qnli\", \"rte\", \"wnli\", \"hans\"]")
361
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets __A = datasets.logging.get_logger(__name__) __A = "\\n@inproceedings{rei-EtAl:2020:WMT,\n author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon},\n title = {Unbabel's Participation in the WMT20 Metrics Shared Task},\n booktitle = {Proceedings of the Fifth Conference on Machine Translation},\n month = {November},\n year = {2020},\n address = {Online},\n publisher = {Association for Computational Linguistics},\n pages = {909--918},\n}\n@inproceedings{rei-etal-2020-comet,\n title = \"{COMET}: A Neural Framework for {MT} Evaluation\",\n author = \"Rei, Ricardo and\n Stewart, Craig and\n Farinha, Ana C and\n Lavie, Alon\",\n booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\",\n month = nov,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\",\n pages = \"2685--2702\",\n}\n" __A = "\\nCrosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA's or MQM).\nWith the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition.\n\nSee the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information.\n" __A = "\nCOMET score.\n\nArgs:\n\n`sources` (list of str): Source sentences\n`predictions` (list of str): candidate translations\n`references` (list of str): reference translations\n`cuda` (bool): If set to True, runs COMET using GPU\n`show_progress` (bool): Shows progress\n`model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None.\n\nReturns:\n `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`.\n `scores`: List of scores.\n\nExamples:\n\n >>> comet_metric = datasets.load_metric('comet')\n >>> # comet_metric = load_metric('comet', 'wmt20-comet-da') # you can also choose which model to use\n >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"]\n >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"]\n >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"]\n >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source)\n >>> print([round(v, 2) for v in results[\"scores\"]])\n [0.19, 0.92]\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class snake_case ( datasets.Metric ): def lowercase_ ( self : List[Any])-> Dict: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://unbabel.github.io/COMET/html/index.html" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "sources": datasets.Value("string" , id="sequence"), "predictions": datasets.Value("string" , id="sequence"), "references": datasets.Value("string" , id="sequence"), }) , codebase_urls=["https://github.com/Unbabel/COMET"] , reference_urls=[ "https://github.com/Unbabel/COMET", "https://www.aclweb.org/anthology/2020.emnlp-main.213/", "http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6", ] , ) def lowercase_ ( self : Tuple , UpperCamelCase__ : Any)-> Dict: '''simple docstring''' if self.config_name == "default": __lowerCAmelCase: Union[str, Any] = comet.load_from_checkpoint(comet.download_model("wmt20-comet-da")) else: __lowerCAmelCase: Tuple = comet.load_from_checkpoint(comet.download_model(self.config_name)) def lowercase_ ( self : Optional[int] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int , UpperCamelCase__ : Optional[int]=None , UpperCamelCase__ : List[str]=False)-> str: '''simple docstring''' if gpus is None: __lowerCAmelCase: Union[str, Any] = 1 if torch.cuda.is_available() else 0 __lowerCAmelCase: Dict = {"src": sources, "mt": predictions, "ref": references} __lowerCAmelCase: Union[str, Any] = [dict(zip(UpperCamelCase__ , UpperCamelCase__)) for t in zip(*data.values())] __lowerCAmelCase , __lowerCAmelCase: str = self.scorer.predict(UpperCamelCase__ , gpus=UpperCamelCase__ , progress_bar=UpperCamelCase__) return {"mean_score": mean_score, "scores": scores}
108
0
import unittest from typing import Dict, List, Optional, Union import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BridgeTowerImageProcessor class __lowerCAmelCase ( unittest.TestCase ): def __init__( self : Any , A : int , A : Any = True , A : int = None , A : Dict = 32 , A : Optional[Any] = True , A : Tuple = 1 / 2_55 , A : str = True , A : List[Any] = True , A : List[Any] = [0.4_8_1_4_5_4_6_6, 0.4_5_7_8_2_7_5, 0.4_0_8_2_1_0_7_3] , A : Tuple = [0.2_6_8_6_2_9_5_4, 0.2_6_1_3_0_2_5_8, 0.2_7_5_7_7_7_1_1] , A : Union[str, Any] = True , A : int=7 , A : int=30 , A : Union[str, Any]=4_00 , A : Tuple=3 , ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = parent _UpperCAmelCase = do_resize _UpperCAmelCase = size if size is not None else {'shortest_edge': 2_88} _UpperCAmelCase = size_divisor _UpperCAmelCase = do_rescale _UpperCAmelCase = rescale_factor _UpperCAmelCase = do_normalize _UpperCAmelCase = do_center_crop _UpperCAmelCase = image_mean _UpperCAmelCase = image_std _UpperCAmelCase = do_pad _UpperCAmelCase = batch_size _UpperCAmelCase = num_channels _UpperCAmelCase = min_resolution _UpperCAmelCase = max_resolution def _lowerCamelCase ( self : List[str]) -> Any: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "size_divisor": self.size_divisor, } def _lowerCamelCase ( self : Any , A : Any , A : str=False) -> str: """simple docstring""" if not batched: _UpperCAmelCase = self.size['shortest_edge'] _UpperCAmelCase = image_inputs[0] if isinstance(_snake_case , Image.Image): _UpperCAmelCase , _UpperCAmelCase = image.size else: _UpperCAmelCase , _UpperCAmelCase = image.shape[1], image.shape[2] _UpperCAmelCase = size / min(_snake_case , _snake_case) if h < w: _UpperCAmelCase , _UpperCAmelCase = size, scale * w else: _UpperCAmelCase , _UpperCAmelCase = scale * h, size _UpperCAmelCase = int((13_33 / 8_00) * size) if max(_snake_case , _snake_case) > max_size: _UpperCAmelCase = max_size / max(_snake_case , _snake_case) _UpperCAmelCase = newh * scale _UpperCAmelCase = neww * scale _UpperCAmelCase , _UpperCAmelCase = int(newh + 0.5), int(neww + 0.5) _UpperCAmelCase , _UpperCAmelCase = ( newh // self.size_divisor * self.size_divisor, neww // self.size_divisor * self.size_divisor, ) else: _UpperCAmelCase = [] for image in image_inputs: _UpperCAmelCase , _UpperCAmelCase = self.get_expected_values([image]) expected_values.append((expected_height, expected_width)) _UpperCAmelCase = max(_snake_case , key=lambda A: item[0])[0] _UpperCAmelCase = max(_snake_case , key=lambda A: item[1])[1] return expected_height, expected_width @require_torch @require_vision class __lowerCAmelCase ( A , unittest.TestCase ): UpperCamelCase = BridgeTowerImageProcessor if is_vision_available() else None def _lowerCamelCase ( self : Tuple) -> Any: """simple docstring""" _UpperCAmelCase = BridgeTowerImageProcessingTester(self) @property def _lowerCamelCase ( self : Dict) -> List[str]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowerCamelCase ( self : List[str]) -> List[Any]: """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(_snake_case , 'image_mean')) self.assertTrue(hasattr(_snake_case , 'image_std')) self.assertTrue(hasattr(_snake_case , 'do_normalize')) self.assertTrue(hasattr(_snake_case , 'do_resize')) self.assertTrue(hasattr(_snake_case , 'size')) self.assertTrue(hasattr(_snake_case , 'size_divisor')) def _lowerCamelCase ( self : List[Any]) -> str: """simple docstring""" pass def _lowerCamelCase ( self : Tuple) -> List[str]: """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict) # create random PIL images _UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case) for image in image_inputs: self.assertIsInstance(_snake_case , Image.Image) # Test not batched input _UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='pt').pixel_values _UpperCAmelCase , _UpperCAmelCase = self.image_processor_tester.get_expected_values(_snake_case) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _UpperCAmelCase = image_processing(_snake_case , return_tensors='pt').pixel_values _UpperCAmelCase , _UpperCAmelCase = self.image_processor_tester.get_expected_values(_snake_case , batched=_snake_case) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _lowerCamelCase ( self : Tuple) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors _UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , numpify=_snake_case) for image in image_inputs: self.assertIsInstance(_snake_case , np.ndarray) # Test not batched input _UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='pt').pixel_values _UpperCAmelCase , _UpperCAmelCase = self.image_processor_tester.get_expected_values(_snake_case) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _UpperCAmelCase = image_processing(_snake_case , return_tensors='pt').pixel_values _UpperCAmelCase , _UpperCAmelCase = self.image_processor_tester.get_expected_values(_snake_case , batched=_snake_case) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _lowerCamelCase ( self : List[str]) -> int: """simple docstring""" _UpperCAmelCase = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors _UpperCAmelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=_snake_case , torchify=_snake_case) for image in image_inputs: self.assertIsInstance(_snake_case , torch.Tensor) # Test not batched input _UpperCAmelCase = image_processing(image_inputs[0] , return_tensors='pt').pixel_values _UpperCAmelCase , _UpperCAmelCase = self.image_processor_tester.get_expected_values(_snake_case) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _UpperCAmelCase = image_processing(_snake_case , return_tensors='pt').pixel_values _UpperCAmelCase , _UpperCAmelCase = self.image_processor_tester.get_expected_values(_snake_case , batched=_snake_case) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , )
339
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A : Dict = { 'configuration_xlm_roberta': [ 'XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaConfig', 'XLMRobertaOnnxConfig', ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Union[str, Any] = ['XLMRobertaTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : int = ['XLMRobertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : List[Any] = [ 'XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMRobertaForCausalLM', 'XLMRobertaForMaskedLM', 'XLMRobertaForMultipleChoice', 'XLMRobertaForQuestionAnswering', 'XLMRobertaForSequenceClassification', 'XLMRobertaForTokenClassification', 'XLMRobertaModel', 'XLMRobertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : int = [ 'TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMRobertaForCausalLM', 'TFXLMRobertaForMaskedLM', 'TFXLMRobertaForMultipleChoice', 'TFXLMRobertaForQuestionAnswering', 'TFXLMRobertaForSequenceClassification', 'TFXLMRobertaForTokenClassification', 'TFXLMRobertaModel', 'TFXLMRobertaPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Tuple = [ 'FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxXLMRobertaForMaskedLM', 'FlaxXLMRobertaForCausalLM', 'FlaxXLMRobertaForMultipleChoice', 'FlaxXLMRobertaForQuestionAnswering', 'FlaxXLMRobertaForSequenceClassification', 'FlaxXLMRobertaForTokenClassification', 'FlaxXLMRobertaModel', 'FlaxXLMRobertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys A : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
6
0
import numpy as np def snake_case ( snake_case__ :np.ndarray , snake_case__ :np.ndarray , snake_case__ :float = 1E-12 , snake_case__ :int = 100 , ) -> tuple[float, np.ndarray]: """simple docstring""" assert np.shape(snake_case__)[0] == np.shape(snake_case__)[1] # Ensure proper dimensionality. assert np.shape(snake_case__)[0] == np.shape(snake_case__)[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(snake_case__) == np.iscomplexobj(snake_case__) _A = np.iscomplexobj(snake_case__) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(snake_case__ , input_matrix.conj().T) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. _A = False _A = 0 _A = 0 _A = 1E12 while not convergence: # Multiple matrix by the vector. _A = np.dot(snake_case__ , snake_case__) # Normalize the resulting output vector. _A = w / np.linalg.norm(snake_case__) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) _A = vector.conj().T if is_complex else vector.T _A = np.dot(snake_case__ , np.dot(snake_case__ , snake_case__)) # Check convergence. _A = np.abs(lambda_ - lambda_previous) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: _A = True _A = lambda_ if is_complex: _A = np.real(lambda_) return lambda_, vector def snake_case ( ) -> None: """simple docstring""" _A = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]]) _A = np.array([41, 4, 20]) _A = real_input_matrix.astype(np.complexaaa) _A = np.triu(1j * complex_input_matrix , 1) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T _A = np.array([41, 4, 20]).astype(np.complexaaa) for problem_type in ["real", "complex"]: if problem_type == "real": _A = real_input_matrix _A = real_vector elif problem_type == "complex": _A = complex_input_matrix _A = complex_vector # Our implementation. _A , _A = power_iteration(snake_case__ , snake_case__) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). _A , _A = np.linalg.eigh(snake_case__) # Last eigenvalue is the maximum one. _A = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. _A = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max) <= 1E-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(snake_case__) - np.abs(snake_case__)) <= 1E-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
357
import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup _SCREAMING_SNAKE_CASE = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36' ' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582' } def snake_case ( snake_case__ :str = "dhaka" , snake_case__ :int = 5) -> int: _A = min(snake_case__ , 50) # Prevent abuse! _A = { """q""": query, """tbm""": """isch""", """hl""": """en""", """ijn""": """0""", } _A = requests.get("""https://www.google.com/search""" , params=snake_case__ , headers=snake_case__) _A = BeautifulSoup(html.text , """html.parser""") _A = """""".join( re.findall(R"""AF_initDataCallback\(([^<]+)\);""" , str(soup.select("""script""")))) _A = json.dumps(snake_case__) _A = json.loads(snake_case__) _A = re.findall( R"""\[\"GRID_STATE0\",null,\[\[1,\[0,\".*?\",(.*),\"All\",""" , snake_case__ , ) if not matched_google_image_data: return 0 _A = re.sub( R"""\[\"(https\:\/\/encrypted-tbn0\.gstatic\.com\/images\?.*?)\",\d+,\d+\]""" , """""" , str(snake_case__) , ) _A = re.findall( R"""(?:'|,),\[\"(https:|http.*?)\",\d+,\d+\]""" , snake_case__ , ) for index, fixed_full_res_image in enumerate(snake_case__): if index >= max_images: return index _A = bytes(snake_case__ , """ascii""").decode( """unicode-escape""") _A = bytes(snake_case__ , """ascii""").decode( """unicode-escape""") _A = urllib.request.build_opener() _A = [ ( """User-Agent""", """Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36""" """ (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582""", ) ] urllib.request.install_opener(snake_case__) _A = F'''query_{query.replace(' ' , '_')}''' if not os.path.exists(snake_case__): os.makedirs(snake_case__) urllib.request.urlretrieve( # noqa: S310 snake_case__ , F'''{path_name}/original_size_img_{index}.jpg''') return index if __name__ == "__main__": try: _SCREAMING_SNAKE_CASE = download_images_from_google_query(sys.argv[1]) print(F'''{image_count} images were downloaded to disk.''') except IndexError: print('Please provide a search term.') raise
81
0
'''simple docstring''' from ....configuration_utils import PretrainedConfig from ....utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { "CarlCochet/trajectory-transformer-halfcheetah-medium-v2": ( "https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-medium-v2/resolve/main/config.json" ), # See all TrajectoryTransformer models at https://huggingface.co/models?filter=trajectory_transformer } class A_ ( _UpperCAmelCase ): '''simple docstring''' UpperCAmelCase_ : Any = """trajectory_transformer""" UpperCAmelCase_ : List[str] = ["""past_key_values"""] UpperCAmelCase_ : Optional[int] = { """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Union[str, Any] , lowercase_ : Dict=100 , lowercase_ : Union[str, Any]=5 , lowercase_ : Dict=1 , lowercase_ : List[Any]=1 , lowercase_ : Union[str, Any]=249 , lowercase_ : Union[str, Any]=6 , lowercase_ : Union[str, Any]=17 , lowercase_ : List[str]=25 , lowercase_ : str=4 , lowercase_ : Dict=4 , lowercase_ : List[Any]=128 , lowercase_ : Optional[int]=0.1 , lowercase_ : Union[str, Any]=0.1 , lowercase_ : List[str]=0.1 , lowercase_ : Tuple=0.0006 , lowercase_ : int=512 , lowercase_ : int=0.02 , lowercase_ : Optional[Any]=1E-12 , lowercase_ : Any=1 , lowercase_ : Optional[int]=True , lowercase_ : Optional[int]=1 , lowercase_ : Union[str, Any]=50_256 , lowercase_ : Optional[Any]=50_256 , **lowercase_ : Tuple , ) -> int: UpperCAmelCase : Optional[Any] = vocab_size UpperCAmelCase : int = action_weight UpperCAmelCase : Tuple = reward_weight UpperCAmelCase : str = value_weight UpperCAmelCase : List[Any] = max_position_embeddings UpperCAmelCase : Optional[int] = block_size UpperCAmelCase : List[Any] = action_dim UpperCAmelCase : Union[str, Any] = observation_dim UpperCAmelCase : int = transition_dim UpperCAmelCase : Union[str, Any] = learning_rate UpperCAmelCase : Any = n_layer UpperCAmelCase : List[str] = n_head UpperCAmelCase : Any = n_embd UpperCAmelCase : Dict = embd_pdrop UpperCAmelCase : Any = attn_pdrop UpperCAmelCase : List[str] = resid_pdrop UpperCAmelCase : List[Any] = initializer_range UpperCAmelCase : int = layer_norm_eps UpperCAmelCase : List[str] = kaiming_initializer_range UpperCAmelCase : List[Any] = use_cache super().__init__(pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ , **UpperCAmelCase_ )
151
import random def _lowercase ( UpperCamelCase_ , UpperCamelCase_ ) -> tuple: '''simple docstring''' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = [], [], [] for element in data: if element < pivot: less.append(UpperCamelCase_ ) elif element > pivot: greater.append(UpperCamelCase_ ) else: equal.append(UpperCamelCase_ ) return less, equal, greater def _lowercase ( UpperCamelCase_ , UpperCamelCase_ ) -> Any: '''simple docstring''' if index >= len(UpperCamelCase_ ) or index < 0: return None SCREAMING_SNAKE_CASE__ = items[random.randint(0 , len(UpperCamelCase_ ) - 1 )] SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = _partition(UpperCamelCase_ , UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = len(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = len(UpperCamelCase_ ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(UpperCamelCase_ , UpperCamelCase_ ) # must be in larger else: return quick_select(UpperCamelCase_ , index - (m + count) )
176
0
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): __magic_name__ = StableDiffusionSAGPipeline __magic_name__ = TEXT_TO_IMAGE_PARAMS __magic_name__ = TEXT_TO_IMAGE_BATCH_PARAMS __magic_name__ = TEXT_TO_IMAGE_IMAGE_PARAMS __magic_name__ = TEXT_TO_IMAGE_IMAGE_PARAMS __magic_name__ = False def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" torch.manual_seed(0 ) A : Union[str, Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) A : Tuple = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=_snake_case , set_alpha_to_one=_snake_case , ) torch.manual_seed(0 ) A : Any = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) A : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) A : Union[str, Any] = CLIPTextModel(_snake_case ) A : Dict = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) A : Optional[int] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=0 ) -> Dict: """simple docstring""" if str(_snake_case ).startswith('''mps''' ): A : Dict = torch.manual_seed(_snake_case ) else: A : List[Any] = torch.Generator(device=_snake_case ).manual_seed(_snake_case ) A : Dict = { "prompt": ".", "generator": generator, "num_inference_steps": 2, "guidance_scale": 1.0, "sag_scale": 1.0, "output_type": "numpy", } return inputs def __lowerCAmelCase ( self ) -> int: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class A ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" A : Dict = StableDiffusionSAGPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''' ) A : Tuple = sag_pipe.to(_snake_case ) sag_pipe.set_progress_bar_config(disable=_snake_case ) A : Optional[int] = "." A : Optional[int] = torch.manual_seed(0 ) A : List[str] = sag_pipe( [prompt] , generator=_snake_case , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type='''np''' ) A : Tuple = output.images A : int = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) A : List[Any] = np.array([0.1_568, 0.1_738, 0.1_695, 0.1_693, 0.1_507, 0.1_705, 0.1_547, 0.1_751, 0.1_949] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-2 def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : Optional[int] = StableDiffusionSAGPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''' ) A : Optional[Any] = sag_pipe.to(_snake_case ) sag_pipe.set_progress_bar_config(disable=_snake_case ) A : Union[str, Any] = "." A : Dict = torch.manual_seed(0 ) A : Union[str, Any] = sag_pipe( [prompt] , generator=_snake_case , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type='''np''' ) A : Optional[Any] = output.images A : Optional[int] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) A : Union[str, Any] = np.array([0.3_459, 0.2_876, 0.2_537, 0.3_002, 0.2_671, 0.2_160, 0.3_026, 0.2_262, 0.2_371] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-2 def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : Dict = StableDiffusionSAGPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''' ) A : str = sag_pipe.to(_snake_case ) sag_pipe.set_progress_bar_config(disable=_snake_case ) A : str = "." A : Union[str, Any] = torch.manual_seed(0 ) A : int = sag_pipe( [prompt] , width=768 , height=512 , generator=_snake_case , guidance_scale=7.5 , sag_scale=1.0 , num_inference_steps=20 , output_type='''np''' , ) A : Union[str, Any] = output.images assert image.shape == (1, 512, 768, 3)
351
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : int = { 'asapp/sew-tiny-100k': 'https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json', # See all SEW models at https://huggingface.co/models?filter=sew } class A ( __snake_case ): __magic_name__ = '''sew''' def __init__( self , SCREAMING_SNAKE_CASE=32 , SCREAMING_SNAKE_CASE=768 , SCREAMING_SNAKE_CASE=12 , SCREAMING_SNAKE_CASE=12 , SCREAMING_SNAKE_CASE=3072 , SCREAMING_SNAKE_CASE=2 , SCREAMING_SNAKE_CASE="gelu" , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.0 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.02 , SCREAMING_SNAKE_CASE=1e-5 , SCREAMING_SNAKE_CASE="group" , SCREAMING_SNAKE_CASE="gelu" , SCREAMING_SNAKE_CASE=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , SCREAMING_SNAKE_CASE=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , SCREAMING_SNAKE_CASE=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=128 , SCREAMING_SNAKE_CASE=16 , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=0.05 , SCREAMING_SNAKE_CASE=10 , SCREAMING_SNAKE_CASE=2 , SCREAMING_SNAKE_CASE=0.0 , SCREAMING_SNAKE_CASE=10 , SCREAMING_SNAKE_CASE=0 , SCREAMING_SNAKE_CASE="mean" , SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=256 , SCREAMING_SNAKE_CASE=0 , SCREAMING_SNAKE_CASE=1 , SCREAMING_SNAKE_CASE=2 , **SCREAMING_SNAKE_CASE , ) -> Tuple: """simple docstring""" super().__init__(**SCREAMING_SNAKE_CASE , pad_token_id=SCREAMING_SNAKE_CASE , bos_token_id=SCREAMING_SNAKE_CASE , eos_token_id=SCREAMING_SNAKE_CASE ) A : Optional[Any] = hidden_size A : Any = feat_extract_norm A : Optional[int] = feat_extract_activation A : Tuple = list(SCREAMING_SNAKE_CASE ) A : List[str] = list(SCREAMING_SNAKE_CASE ) A : List[str] = list(SCREAMING_SNAKE_CASE ) A : int = conv_bias A : List[Any] = num_conv_pos_embeddings A : Tuple = num_conv_pos_embedding_groups A : int = len(self.conv_dim ) A : Dict = num_hidden_layers A : Optional[int] = intermediate_size A : Any = squeeze_factor A : int = hidden_act A : str = num_attention_heads A : Dict = hidden_dropout A : Optional[Any] = attention_dropout A : List[str] = activation_dropout A : Union[str, Any] = feat_proj_dropout A : Union[str, Any] = final_dropout A : int = layerdrop A : Optional[Any] = layer_norm_eps A : Any = initializer_range A : Tuple = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' F'but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)' F'= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 A : Optional[Any] = apply_spec_augment A : Optional[Any] = mask_time_prob A : Union[str, Any] = mask_time_length A : Optional[Any] = mask_time_min_masks A : str = mask_feature_prob A : Tuple = mask_feature_length A : Any = mask_feature_min_masks # ctc loss A : List[Any] = ctc_loss_reduction A : Dict = ctc_zero_infinity # sequence classification A : int = use_weighted_layer_sum A : Optional[int] = classifier_proj_size @property def __lowerCAmelCase ( self ) -> Optional[Any]: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
311
0
"""simple docstring""" def UpperCamelCase__ ( lowercase__ : list , lowercase__ : int , lowercase__ : int = 0 , lowercase__ : int = 0 ): snake_case : int = right or len(lowercase__ ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(lowercase__ , lowercase__ , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
148
"""simple docstring""" import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder __A = "base_with_context" def UpperCamelCase__ ( lowercase__ : Optional[Any] , lowercase__ : List[Any] ): snake_case : Dict = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=lowercase__ ) for lyr_num, lyr in enumerate(model.encoders ): snake_case : Tuple = weights[F'''layers_{lyr_num}'''] snake_case : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) snake_case : List[Any] = ly_weight["attention"] snake_case : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : Any = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def UpperCamelCase__ ( lowercase__ : Tuple , lowercase__ : List[Any] ): snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=lowercase__ ) for lyr_num, lyr in enumerate(model.encoders ): snake_case : str = weights[F'''layers_{lyr_num}'''] snake_case : Any = ly_weight["attention"] snake_case : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : int = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) snake_case : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) snake_case : int = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def UpperCamelCase__ ( lowercase__ : str , lowercase__ : Union[str, Any] ): snake_case : int = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ) , requires_grad=lowercase__ ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): snake_case : Union[str, Any] = weights[F'''layers_{lyr_num}'''] snake_case : List[str] = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) snake_case : Any = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) snake_case : Union[str, Any] = ly_weight["self_attention"] snake_case : List[str] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : Any = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : List[str] = ly_weight["MultiHeadDotProductAttention_0"] snake_case : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) snake_case : Union[str, Any] = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) snake_case : Dict = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) snake_case : Tuple = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) snake_case : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) snake_case : Any = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) snake_case : str = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) snake_case : List[str] = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) snake_case : List[Any] = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def UpperCamelCase__ ( lowercase__ : Any ): snake_case : Union[str, Any] = checkpoints.load_tax_checkpoint(args.checkpoint_path ) snake_case : List[Any] = jnp.tree_util.tree_map(onp.array , lowercase__ ) snake_case : Tuple = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] snake_case : List[str] = os.path.join(args.checkpoint_path , ".." , "config.gin" ) snake_case : List[str] = inference.parse_training_gin_file(lowercase__ , lowercase__ ) snake_case : List[Any] = inference.InferenceModel(args.checkpoint_path , lowercase__ ) snake_case : str = DDPMScheduler(beta_schedule="squaredcos_cap_v2" , variance_type="fixed_large" ) snake_case : int = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) snake_case : Tuple = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length["targets_context"] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) snake_case : str = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length["targets_context"] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) snake_case : Optional[int] = load_notes_encoder(ta_checkpoint["target"]["token_encoder"] , lowercase__ ) snake_case : Any = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"] , lowercase__ ) snake_case : List[Any] = load_decoder(ta_checkpoint["target"]["decoder"] , lowercase__ ) snake_case : int = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) snake_case : Tuple = SpectrogramDiffusionPipeline( notes_encoder=lowercase__ , continuous_encoder=lowercase__ , decoder=lowercase__ , scheduler=lowercase__ , melgan=lowercase__ , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": __A = argparse.ArgumentParser() parser.add_argument("--output_path", default=None, type=str, required=True, help="Path to the converted model.") parser.add_argument( "--save", default=True, type=bool, required=False, help="Whether to save the converted model or not." ) parser.add_argument( "--checkpoint_path", default=f'{MODEL}/checkpoint_500000', type=str, required=False, help="Path to the original jax model checkpoint.", ) __A = parser.parse_args() main(args)
148
1
"""simple docstring""" import random from typing import Any def _A ( UpperCamelCase_ : list) -> list[Any]: '''simple docstring''' for _ in range(len(UpperCamelCase_)): __lowercase = random.randint(0, len(UpperCamelCase_) - 1) __lowercase = random.randint(0, len(UpperCamelCase_) - 1) __lowercase ,__lowercase = data[b], data[a] return data if __name__ == "__main__": _a = [0, 1, 2, 3, 4, 5, 6, 7] _a = ['python', 'says', 'hello', '!'] print('Fisher-Yates Shuffle:') print('List', integers, strings) print('FY Shuffle', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
144
"""simple docstring""" import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def _A ( ) -> None: '''simple docstring''' print("Making key files...") make_key_files("rsa", 1024) print("Key files generation successful.") def _A ( UpperCamelCase_ : int) -> tuple[tuple[int, int], tuple[int, int]]: '''simple docstring''' print("Generating prime p...") __lowercase = rabinMiller.generate_large_prime(UpperCamelCase_) print("Generating prime q...") __lowercase = rabinMiller.generate_large_prime(UpperCamelCase_) __lowercase = p * q print("Generating e that is relatively prime to (p - 1) * (q - 1)...") while True: __lowercase = random.randrange(2 ** (key_size - 1), 2 ** (key_size)) if cryptoMath.gcd(UpperCamelCase_, (p - 1) * (q - 1)) == 1: break print("Calculating d that is mod inverse of e...") __lowercase = cryptoMath.find_mod_inverse(UpperCamelCase_, (p - 1) * (q - 1)) __lowercase = (n, e) __lowercase = (n, d) return (public_key, private_key) def _A ( UpperCamelCase_ : str, UpperCamelCase_ : int) -> None: '''simple docstring''' if os.path.exists(F"""{name}_pubkey.txt""") or os.path.exists(F"""{name}_privkey.txt"""): print("\nWARNING:") print( F"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" "Use a different name or delete these files and re-run this program.") sys.exit() __lowercase ,__lowercase = generate_key(UpperCamelCase_) print(F"""\nWriting public key to file {name}_pubkey.txt...""") with open(F"""{name}_pubkey.txt""", "w") as out_file: out_file.write(F"""{key_size},{public_key[0]},{public_key[1]}""") print(F"""Writing private key to file {name}_privkey.txt...""") with open(F"""{name}_privkey.txt""", "w") as out_file: out_file.write(F"""{key_size},{private_key[0]},{private_key[1]}""") if __name__ == "__main__": main()
144
1
'''simple docstring''' import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask A__ : Any =logging.getLogger(__name__) class UpperCAmelCase ( A_ ): def __init__( self : Optional[int] , __snake_case : Tuple=-1 ) -> int: _lowerCAmelCase = label_idx def lowercase__ ( self : Any , __snake_case : int , __snake_case : Union[Split, str] ) -> List[InputExample]: if isinstance(_snake_case , _snake_case ): _lowerCAmelCase = mode.value _lowerCAmelCase = os.path.join(_snake_case , f"{mode}.txt" ) _lowerCAmelCase = 1 _lowerCAmelCase = [] with open(_snake_case , encoding="""utf-8""" ) as f: _lowerCAmelCase = [] _lowerCAmelCase = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=f"{mode}-{guid_index}" , words=_snake_case , labels=_snake_case ) ) guid_index += 1 _lowerCAmelCase = [] _lowerCAmelCase = [] else: _lowerCAmelCase = line.split(""" """ ) words.append(splits[0] ) if len(_snake_case ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=f"{mode}-{guid_index}" , words=_snake_case , labels=_snake_case ) ) return examples def lowercase__ ( self : str , __snake_case : TextIO , __snake_case : TextIO , __snake_case : List ) -> List[Any]: _lowerCAmelCase = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(_snake_case ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _lowerCAmelCase = line.split()[0] + ''' ''' + preds_list[example_id].pop(0 ) + '''\n''' writer.write(_snake_case ) else: logger.warning("""Maximum sequence length exceeded: No prediction for \'%s\'.""" , line.split()[0] ) def lowercase__ ( self : Tuple , __snake_case : str ) -> List[str]: if path: with open(_snake_case , """r""" ) as f: _lowerCAmelCase = f.read().splitlines() if "O" not in labels: _lowerCAmelCase = ['''O'''] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class UpperCAmelCase ( A_ ): def __init__( self : str ) -> Any: super().__init__(label_idx=-2 ) def lowercase__ ( self : str , __snake_case : str ) -> List[str]: if path: with open(_snake_case , """r""" ) as f: _lowerCAmelCase = f.read().splitlines() if "O" not in labels: _lowerCAmelCase = ['''O'''] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class UpperCAmelCase ( A_ ): def lowercase__ ( self : int , __snake_case : Optional[int] , __snake_case : Union[Split, str] ) -> List[InputExample]: if isinstance(_snake_case , _snake_case ): _lowerCAmelCase = mode.value _lowerCAmelCase = os.path.join(_snake_case , f"{mode}.txt" ) _lowerCAmelCase = 1 _lowerCAmelCase = [] with open(_snake_case , encoding="""utf-8""" ) as f: for sentence in parse_incr(_snake_case ): _lowerCAmelCase = [] _lowerCAmelCase = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(_snake_case ) == len(_snake_case ) if words: examples.append(InputExample(guid=f"{mode}-{guid_index}" , words=_snake_case , labels=_snake_case ) ) guid_index += 1 return examples def lowercase__ ( self : List[Any] , __snake_case : TextIO , __snake_case : TextIO , __snake_case : List ) -> str: _lowerCAmelCase = 0 for sentence in parse_incr(_snake_case ): _lowerCAmelCase = preds_list[example_id] _lowerCAmelCase = '''''' for token in sentence: out += f"{token['form']} ({token['upos']}|{s_p.pop(0 )}) " out += "\n" writer.write(_snake_case ) example_id += 1 def lowercase__ ( self : List[str] , __snake_case : str ) -> List[str]: if path: with open(_snake_case , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
70
"""simple docstring""" import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCAmelCase_ = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCAmelCase_ = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCAmelCase_ = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCAmelCase_ = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): '''simple docstring''' def UpperCAmelCase ( self : Optional[int] ) -> str: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''string''' ,id='''sequence''' ), '''references''': datasets.Value('''string''' ,id='''sequence''' ), } ) ,codebase_urls=['''https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'''] ,reference_urls=[ '''https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score''', '''https://en.wikipedia.org/wiki/METEOR''', ] ,) def UpperCAmelCase ( self : str ,_snake_case : Dict ) -> Dict: """simple docstring""" import nltk nltk.download('''wordnet''' ) if NLTK_VERSION >= version.Version('''3.6.5''' ): nltk.download('''punkt''' ) if NLTK_VERSION >= version.Version('''3.6.6''' ): nltk.download('''omw-1.4''' ) def UpperCAmelCase ( self : Dict ,_snake_case : Dict ,_snake_case : List[str] ,_snake_case : Tuple=0.9 ,_snake_case : Optional[int]=3 ,_snake_case : Union[str, Any]=0.5 ) -> List[str]: """simple docstring""" if NLTK_VERSION >= version.Version('''3.6.5''' ): lowercase__ : int = [ meteor_score.single_meteor_score( word_tokenize(_snake_case ) ,word_tokenize(_snake_case ) ,alpha=_snake_case ,beta=_snake_case ,gamma=_snake_case ) for ref, pred in zip(_snake_case ,_snake_case ) ] else: lowercase__ : Tuple = [ meteor_score.single_meteor_score(_snake_case ,_snake_case ,alpha=_snake_case ,beta=_snake_case ,gamma=_snake_case ) for ref, pred in zip(_snake_case ,_snake_case ) ] return {"meteor": np.mean(_snake_case )}
16
0
'''simple docstring''' import argparse import json import os from collections import OrderedDict import numpy as np import tensorflow as tf import torch def snake_case__ ( lowerCamelCase__ : Dict ) -> Any: A_ : Union[str, Any] = os.path.join(args.tf_model_dir , '''parameters.json''' ) A_ : Tuple = json.loads(open(__snake_case ).read() ) if not params: raise ValueError( f'It seems that the json file at {parameter_file} is empty. Make sure you have a correct json file.' ) if not args.output.endswith('''.pt''' ): A_ : Dict = args.output + ".pt" A_ : str = OrderedDict() with tf.device('''/CPU:0''' ): A_ : int = tf.train.load_checkpoint(args.tf_model_dir ) A_ : Dict = reader.get_variable_to_shape_map() for key_name in shapes.keys(): A_ : Any = reader.get_tensor(__snake_case ).astype(np.floataa ) if key_name.endswith('''/adam_m''' ) or key_name.endswith('''/adam_v''' ): continue if key_name.startswith('''pasts/''' ): if key_name.startswith('''pasts/mlp''' ): A_ : Optional[Any] = int(key_name[9] ) elif key_name.startswith('''pasts/out''' ): A_ : Tuple = 8 A_ : Optional[int] = "model.sqout.%d.weight" % (player * 2) # enter to nn.Sequencial with Tanh, so 2 at a time A_ : List[str] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix A_ : Tuple = torch.tensor(__snake_case ) elif key_name.startswith('''model/moe''' ): A_ : Union[str, Any] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/switch_gating/kernel''' ): A_ : Tuple = "model.blocks.%d.feed_forward.mlp.router.classifier.weight" % player A_ : Tuple = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix A_ : Optional[Any] = torch.tensor(__snake_case ) elif key_name.endswith('''/softmlp/kernel''' ): A_ : Optional[int] = "model.blocks.%d.feed_forward.soft_bypass_mlp.weight" % player A_ : str = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix A_ : str = torch.tensor(__snake_case ) elif key_name.endswith('''/wo/kernel''' ) or key_name.endswith('''/wi/kernel''' ): A_ : Tuple = key_name[-9:-7] for i in range(1_6 ): A_ : Optional[int] = "model.blocks.%d.feed_forward.mlp.experts.expert_%d.%s.weight" % (player, i, nlayer) A_ : List[Any] = ( vnp[i].transpose([1, 0] ).copy() ) # In Mesh-Tensorflow, it is one array, so it is divided A_ : Dict = torch.tensor(__snake_case ) elif key_name.startswith('''model/mlp''' ): A_ : Optional[int] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/p1/kernel''' ): A_ : Any = "model.blocks.%d.feed_forward.mlp.wi.weight" % player A_ : Optional[int] = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix A_ : Dict = torch.tensor(__snake_case ) elif key_name.endswith('''/p1/bias''' ): A_ : Tuple = "model.blocks.%d.feed_forward.mlp.wi.bias" % player A_ : Optional[Any] = vnp.copy() # same because it is one dimensional A_ : Optional[Any] = torch.tensor(__snake_case ) elif key_name.endswith('''/p2/kernel''' ): A_ : List[Any] = "model.blocks.%d.feed_forward.mlp.wo.weight" % player A_ : Any = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix A_ : Tuple = torch.tensor(__snake_case ) elif key_name.endswith('''/p2/bias''' ): A_ : List[Any] = "model.blocks.%d.feed_forward.mlp.wo.bias" % player A_ : Union[str, Any] = vnp.copy() # same because it is one dimensional A_ : Tuple = torch.tensor(__snake_case ) elif key_name.startswith('''model/ln''' ): A_ : Optional[Any] = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): A_ : List[str] = "model.blocks.%d.feed_forward.norm.bias" % player A_ : List[str] = vnp.copy() # same because it is one dimensional A_ : Tuple = torch.tensor(__snake_case ) elif key_name.endswith('''/g''' ): A_ : Optional[int] = "model.blocks.%d.feed_forward.norm.weight" % player A_ : Optional[int] = vnp.copy() # same because it is one dimensional A_ : List[str] = torch.tensor(__snake_case ) elif key_name.startswith('''model/att''' ): A_ : Optional[Any] = int(key_name[9:].split('''/''' )[0] ) if key_name.endswith('''/qkv/kernel''' ): A_ : int = vnp.copy() # Compute same dimension as Mesh-tensorflow using einsum A_ : str = state[:, 0, :, :] A_ : Union[str, Any] = state[:, 1, :, :] A_ : List[Any] = state[:, 2, :, :] A_ : List[Any] = ( state_q.reshape([state_q.shape[0], state_q.shape[1] * state_q.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix A_ : str = ( state_k.reshape([state_k.shape[0], state_k.shape[1] * state_k.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix A_ : int = ( state_v.reshape([state_v.shape[0], state_v.shape[1] * state_v.shape[2]] ) .transpose([1, 0] ) .copy() ) # Mesh-Tensorflow is a diagonal matrix A_ : Dict = "model.blocks.%d.self_attn.self_attn.q_proj.weight" % player A_ : Tuple = torch.tensor(__snake_case ) A_ : Union[str, Any] = "model.blocks.%d.self_attn.self_attn.k_proj.weight" % player A_ : str = torch.tensor(__snake_case ) A_ : int = "model.blocks.%d.self_attn.self_attn.v_proj.weight" % player A_ : Union[str, Any] = torch.tensor(__snake_case ) elif key_name.endswith('''/o/kernel''' ): A_ : Tuple = "model.blocks.%d.self_attn.self_attn.out_proj.weight" % player A_ : List[str] = ( vnp.reshape([vnp.shape[0] * vnp.shape[1], vnp.shape[2]] ).transpose([1, 0] ).copy() ) # Mesh-Tensorflow is a diagonal matrix A_ : str = torch.tensor(__snake_case ) elif key_name.startswith('''model/an''' ): A_ : Dict = int(key_name[8:].split('''/''' )[0] ) if key_name.endswith('''/b''' ): A_ : str = "model.blocks.%d.self_attn.norm.bias" % player A_ : str = vnp.copy() # same because it is one dimensional A_ : Optional[int] = torch.tensor(__snake_case ) elif key_name.endswith('''/g''' ): A_ : str = "model.blocks.%d.self_attn.norm.weight" % player A_ : Dict = vnp.copy() # same because it is one dimensional A_ : List[Any] = torch.tensor(__snake_case ) elif ( key_name.startswith('''model/wte''' ) or key_name.startswith('''model/wpe''' ) or key_name.startswith('''model/ete''' ) ): A_ : Optional[Any] = {"wte": "embed_tokens", "wpe": "position_embeddings", "ete": "extra_position_embeddings"}[ key_name[-3:] ] A_ : Dict = "model.%s.weight" % nlayer A_ : List[Any] = vnp.copy() # same in embedded A_ : Tuple = torch.tensor(__snake_case ) if key_name.startswith('''model/wte''' ): A_ : Any = "lm_head.weight" A_ : Optional[int] = vnp.copy() # same in embedded A_ : str = torch.tensor(__snake_case ) elif key_name.startswith('''model/wob''' ): A_ : Any = "final_logits_bias" A_ : Union[str, Any] = vnp.copy() # same in embedded A_ : Dict = state.reshape((1, -1) ) A_ : int = torch.tensor(__snake_case ) elif key_name == "model/dense/kernel": A_ : Optional[int] = "model.last_project.weight" A_ : Dict = vnp.transpose([1, 0] ).copy() # Mesh-Tensorflow is a diagonal matrix A_ : List[Any] = torch.tensor(__snake_case ) elif key_name == "model/dense_1/bias": A_ : Optional[int] = "model.last_project.bias" A_ : Tuple = vnp.copy() # same because it is one dimensional A_ : Tuple = torch.tensor(__snake_case ) torch.save(__snake_case , args.output ) if __name__ == "__main__": snake_case__ = argparse.ArgumentParser( description="""model converter.""", formatter_class=argparse.ArgumentDefaultsHelpFormatter ) parser.add_argument("""--tf_model_dir""", metavar="""PATH""", type=str, required=True, help="""import model""") parser.add_argument("""--output""", metavar="""PATH""", type=str, required=True, help="""output model""") snake_case__ = parser.parse_args() convert_tf_gptsan_to_pt(args)
363
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case__ = logging.get_logger(__name__) snake_case__ = { """facebook/s2t-wav2vec2-large-en-de""": ( """https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json""" ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = 'speech_to_text_2' _lowerCAmelCase = ['past_key_values'] _lowerCAmelCase = {'num_attention_heads': 'decoder_attention_heads', 'hidden_size': 'd_model'} def __init__( self : Optional[Any] , _lowerCamelCase : Optional[Any]=10000 , _lowerCamelCase : List[Any]=6 , _lowerCamelCase : int=2048 , _lowerCamelCase : Dict=4 , _lowerCamelCase : str=0.0 , _lowerCamelCase : int=True , _lowerCamelCase : int="relu" , _lowerCamelCase : Any=256 , _lowerCamelCase : List[Any]=0.1 , _lowerCamelCase : Tuple=0.0 , _lowerCamelCase : Union[str, Any]=0.0 , _lowerCamelCase : Optional[Any]=0.02 , _lowerCamelCase : int=2 , _lowerCamelCase : List[str]=True , _lowerCamelCase : str=1 , _lowerCamelCase : List[Any]=0 , _lowerCamelCase : Optional[int]=2 , _lowerCamelCase : Tuple=1024 , **_lowerCamelCase : int , ): """simple docstring""" A_ : Optional[int] = vocab_size A_ : Tuple = d_model A_ : List[str] = decoder_ffn_dim A_ : str = decoder_layers A_ : Any = decoder_attention_heads A_ : int = dropout A_ : str = attention_dropout A_ : Optional[int] = activation_dropout A_ : str = activation_function A_ : List[Any] = init_std A_ : Union[str, Any] = decoder_layerdrop A_ : Any = use_cache A_ : Optional[Any] = decoder_layers A_ : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True A_ : Optional[Any] = max_target_positions super().__init__( pad_token_id=_lowerCamelCase , bos_token_id=_lowerCamelCase , eos_token_id=_lowerCamelCase , decoder_start_token_id=_lowerCamelCase , **_lowerCamelCase , )
4
0
'''simple docstring''' import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class a_ ( unittest.TestCase ): '''simple docstring''' def snake_case_( self ) -> List[Any]: _SCREAMING_SNAKE_CASE = """laion/clap-htsat-unfused""" _SCREAMING_SNAKE_CASE = tempfile.mkdtemp() def snake_case_( self , **A ) -> int: return RobertaTokenizer.from_pretrained(self.checkpoint , **A ) def snake_case_( self , **A ) -> Optional[int]: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **A ) def snake_case_( self ) -> Dict: shutil.rmtree(self.tmpdirname ) def snake_case_( self ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = self.get_feature_extractor() _SCREAMING_SNAKE_CASE = ClapProcessor(tokenizer=A , feature_extractor=A ) processor.save_pretrained(self.tmpdirname ) _SCREAMING_SNAKE_CASE = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A ) def snake_case_( self ) -> List[str]: _SCREAMING_SNAKE_CASE = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) _SCREAMING_SNAKE_CASE = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) _SCREAMING_SNAKE_CASE = self.get_feature_extractor(do_normalize=A , padding_value=1.0 ) _SCREAMING_SNAKE_CASE = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=A , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , A ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , A ) def snake_case_( self ) -> List[Any]: _SCREAMING_SNAKE_CASE = self.get_feature_extractor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = ClapProcessor(tokenizer=A , feature_extractor=A ) _SCREAMING_SNAKE_CASE = floats_list((3, 1000) ) _SCREAMING_SNAKE_CASE = feature_extractor(A , return_tensors="""np""" ) _SCREAMING_SNAKE_CASE = processor(audios=A , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def snake_case_( self ) -> str: _SCREAMING_SNAKE_CASE = self.get_feature_extractor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = ClapProcessor(tokenizer=A , feature_extractor=A ) _SCREAMING_SNAKE_CASE = """This is a test string""" _SCREAMING_SNAKE_CASE = processor(text=A ) _SCREAMING_SNAKE_CASE = tokenizer(A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def snake_case_( self ) -> Any: _SCREAMING_SNAKE_CASE = self.get_feature_extractor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = ClapProcessor(tokenizer=A , feature_extractor=A ) _SCREAMING_SNAKE_CASE = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _SCREAMING_SNAKE_CASE = processor.batch_decode(A ) _SCREAMING_SNAKE_CASE = tokenizer.batch_decode(A ) self.assertListEqual(A , A ) def snake_case_( self ) -> Tuple: _SCREAMING_SNAKE_CASE = self.get_feature_extractor() _SCREAMING_SNAKE_CASE = self.get_tokenizer() _SCREAMING_SNAKE_CASE = ClapProcessor(tokenizer=A , feature_extractor=A ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , )
58
import colorsys from PIL import Image # type: ignore def __UpperCamelCase ( lowerCAmelCase__ : float , lowerCAmelCase__ : float , lowerCAmelCase__ : int ): __a : Any = x __a : List[Any] = y for step in range(lowerCAmelCase__ ): # noqa: B007 __a : List[Any] = a * a - b * b + x __a : Tuple = 2 * a * b + y __a : Optional[int] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def __UpperCamelCase ( lowerCAmelCase__ : float ): if distance == 1: return (0, 0, 0) else: return (2_5_5, 2_5_5, 2_5_5) def __UpperCamelCase ( lowerCAmelCase__ : float ): if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_5_5 ) for i in colorsys.hsv_to_rgb(lowerCAmelCase__ , 1 , 1 ) ) def __UpperCamelCase ( lowerCAmelCase__ : int = 8_0_0 , lowerCAmelCase__ : int = 6_0_0 , lowerCAmelCase__ : float = -0.6 , lowerCAmelCase__ : float = 0 , lowerCAmelCase__ : float = 3.2 , lowerCAmelCase__ : int = 5_0 , lowerCAmelCase__ : bool = True , ): __a : int = Image.new('''RGB''' , (image_width, image_height) ) __a : Dict = img.load() # loop through the image-coordinates for image_x in range(lowerCAmelCase__ ): for image_y in range(lowerCAmelCase__ ): # determine the figure-coordinates based on the image-coordinates __a : Optional[Any] = figure_width / image_width * image_height __a : str = figure_center_x + (image_x / image_width - 0.5) * figure_width __a : str = figure_center_y + (image_y / image_height - 0.5) * figure_height __a : Tuple = get_distance(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: __a : Optional[int] = get_color_coded_rgb(lowerCAmelCase__ ) else: __a : Optional[Any] = get_black_and_white_rgb(lowerCAmelCase__ ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure lowercase__ =get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
216
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = '''▁''' lowerCamelCase__ = {'''vocab_file''': '''sentencepiece.bpe.model'''} lowerCamelCase__ = { '''vocab_file''': { '''facebook/mbart-large-50-one-to-many-mmt''': ( '''https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt/resolve/main/sentencepiece.bpe.model''' ), } } lowerCamelCase__ = { '''facebook/mbart-large-50-one-to-many-mmt''': 1024, } # fmt: off lowerCamelCase__ = ['''ar_AR''', '''cs_CZ''', '''de_DE''', '''en_XX''', '''es_XX''', '''et_EE''', '''fi_FI''', '''fr_XX''', '''gu_IN''', '''hi_IN''', '''it_IT''', '''ja_XX''', '''kk_KZ''', '''ko_KR''', '''lt_LT''', '''lv_LV''', '''my_MM''', '''ne_NP''', '''nl_XX''', '''ro_RO''', '''ru_RU''', '''si_LK''', '''tr_TR''', '''vi_VN''', '''zh_CN''', '''af_ZA''', '''az_AZ''', '''bn_IN''', '''fa_IR''', '''he_IL''', '''hr_HR''', '''id_ID''', '''ka_GE''', '''km_KH''', '''mk_MK''', '''ml_IN''', '''mn_MN''', '''mr_IN''', '''pl_PL''', '''ps_AF''', '''pt_XX''', '''sv_SE''', '''sw_KE''', '''ta_IN''', '''te_IN''', '''th_TH''', '''tl_XX''', '''uk_UA''', '''ur_PK''', '''xh_ZA''', '''gl_ES''', '''sl_SI'''] class _UpperCAmelCase ( lowerCAmelCase ): '''simple docstring''' __A = VOCAB_FILES_NAMES __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = PRETRAINED_VOCAB_FILES_MAP __A = ['''input_ids''', '''attention_mask'''] __A = [] __A = [] def __init__( self : str , lowercase_ : str , lowercase_ : Optional[int]=None , lowercase_ : Any=None , lowercase_ : List[Any]="</s>" , lowercase_ : List[Any]="</s>" , lowercase_ : Tuple="<s>" , lowercase_ : Optional[Any]="<unk>" , lowercase_ : int="<pad>" , lowercase_ : Union[str, Any]="<mask>" , lowercase_ : Optional[Dict[str, Any]] = None , **lowercase_ : Dict , ) -> None: """simple docstring""" _UpperCamelCase = AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_) if isinstance(lowercase_ , lowercase_) else mask_token _UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs _UpperCamelCase = kwargs.get("additional_special_tokens" , []) kwargs["additional_special_tokens"] += [ code for code in FAIRSEQ_LANGUAGE_CODES if code not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=lowercase_ , tgt_lang=lowercase_ , eos_token=lowercase_ , unk_token=lowercase_ , sep_token=lowercase_ , cls_token=lowercase_ , pad_token=lowercase_ , mask_token=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , **lowercase_ , ) _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowercase_)) _UpperCamelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token _UpperCamelCase = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _UpperCamelCase = 1 _UpperCamelCase = len(self.sp_model) _UpperCamelCase = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowercase_) } _UpperCamelCase = {v: k for k, v in self.lang_code_to_id.items()} _UpperCamelCase = len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id) _UpperCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} _UpperCamelCase = src_lang if src_lang is not None else "en_XX" _UpperCamelCase = self.lang_code_to_id[self._src_lang] _UpperCamelCase = tgt_lang self.set_src_lang_special_tokens(self._src_lang) @property def __UpperCAmelCase ( self : int) -> int: """simple docstring""" return len(self.sp_model) + len(self.lang_code_to_id) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self._src_lang @src_lang.setter def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : str) -> None: """simple docstring""" _UpperCamelCase = new_src_lang self.set_src_lang_special_tokens(self._src_lang) def __getstate__( self : List[str]) -> Dict: """simple docstring""" _UpperCamelCase = self.__dict__.copy() _UpperCamelCase = None return state def __setstate__( self : List[str] , lowercase_ : Dict) -> None: """simple docstring""" _UpperCamelCase = d # for backward compatibility if not hasattr(self , "sp_model_kwargs"): _UpperCamelCase = {} _UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def __UpperCAmelCase ( self : List[Any]) -> Dict: """simple docstring""" _UpperCamelCase = {self.convert_ids_to_tokens(lowercase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __UpperCAmelCase ( self : str , lowercase_ : str) -> List[str]: """simple docstring""" return self.sp_model.encode(lowercase_ , out_type=lowercase_) def __UpperCAmelCase ( self : Tuple , lowercase_ : str) -> int: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _UpperCamelCase = self.sp_model.PieceToId(lowercase_) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __UpperCAmelCase ( self : Tuple , lowercase_ : int) -> str: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset) def __UpperCAmelCase ( self : Any , lowercase_ : Optional[int]) -> Dict: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = "" _UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowercase_) + token _UpperCamelCase = True _UpperCamelCase = [] else: current_sub_tokens.append(lowercase_) _UpperCamelCase = False out_string += self.sp_model.decode(lowercase_) return out_string.strip() def __UpperCAmelCase ( self : List[Any] , lowercase_ : str , lowercase_ : Optional[str] = None) -> Tuple[str]: """simple docstring""" if not os.path.isdir(lowercase_): logger.error(f'Vocabulary path ({save_directory}) should be a directory') return _UpperCamelCase = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"]) if os.path.abspath(self.vocab_file) != os.path.abspath(lowercase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowercase_) elif not os.path.isfile(self.vocab_file): with open(lowercase_ , "wb") as fi: _UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowercase_) return (out_vocab_file,) def __UpperCAmelCase ( self : int , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None , lowercase_ : bool = False) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_) _UpperCamelCase = [1] * len(self.prefix_tokens) _UpperCamelCase = [1] * len(self.suffix_tokens) if token_ids_a is None: return prefix_ones + ([0] * len(lowercase_)) + suffix_ones return prefix_ones + ([0] * len(lowercase_)) + ([0] * len(lowercase_)) + suffix_ones def __UpperCAmelCase ( self : List[str] , lowercase_ : List[int] , lowercase_ : Optional[List[int]] = None) -> List[int]: """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __UpperCAmelCase ( self : Optional[int] , lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[str] , lowercase_ : Optional[str] , **lowercase_ : Any) -> Dict: """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model") _UpperCamelCase = src_lang _UpperCamelCase = self(lowercase_ , add_special_tokens=lowercase_ , return_tensors=lowercase_ , **lowercase_) _UpperCamelCase = self.convert_tokens_to_ids(lowercase_) _UpperCamelCase = tgt_lang_id return inputs def __UpperCAmelCase ( self : Any , lowercase_ : List[str] , lowercase_ : str = "en_XX" , lowercase_ : Optional[List[str]] = None , lowercase_ : str = "ro_RO" , **lowercase_ : int , ) -> BatchEncoding: """simple docstring""" _UpperCamelCase = src_lang _UpperCamelCase = tgt_lang return super().prepare_seqaseq_batch(lowercase_ , lowercase_ , **lowercase_) def __UpperCAmelCase ( self : Optional[int]) -> str: """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang) def __UpperCAmelCase ( self : Optional[int]) -> List[str]: """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang) def __UpperCAmelCase ( self : List[Any] , lowercase_ : str) -> None: """simple docstring""" _UpperCamelCase = self.lang_code_to_id[src_lang] _UpperCamelCase = [self.cur_lang_code_id] _UpperCamelCase = [self.eos_token_id] def __UpperCAmelCase ( self : List[Any] , lowercase_ : str) -> None: """simple docstring""" _UpperCamelCase = self.lang_code_to_id[tgt_lang] _UpperCamelCase = [self.cur_lang_code_id] _UpperCamelCase = [self.eos_token_id]
63
import requests from bsa import BeautifulSoup def lowerCAmelCase__ ( a__ = "https://www.worldometers.info/coronavirus" ) ->dict: '''simple docstring''' _UpperCamelCase = BeautifulSoup(requests.get(a__ ).text , "html.parser" ) _UpperCamelCase = soup.findAll("h1" ) _UpperCamelCase = soup.findAll("div" , {"class": "maincounter-number"} ) keys += soup.findAll("span" , {"class": "panel-title"} ) values += soup.findAll("div" , {"class": "number-table-main"} ) return {key.text.strip(): value.text.strip() for key, value in zip(a__ , a__ )} if __name__ == "__main__": print('''\033[1m''' + '''COVID-19 Status of the World''' + '''\033[0m\n''') for key, value in world_covidaa_stats().items(): print(F"{key}\n{value}\n")
63
1