code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features a : Any = logging.get_logger(__name__) a : int = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) a : str = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Model type selected in the list: """ + """, """.join(_UpperCamelCase )} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """The input data dir. Should contain the .json files for the SQuAD task."""} ) __SCREAMING_SNAKE_CASE = field( default=128 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=128 , metadata={"""help""": """When splitting up a long document into chunks, how much stride to take between chunks."""} , ) __SCREAMING_SNAKE_CASE = field( default=64 , metadata={ """help""": ( """The maximum number of tokens for the question. Questions longer than this will """ """be truncated to this length.""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=30 , metadata={ """help""": ( """The maximum length of an answer that can be generated. This is needed because the start """ """and end predictions are not conditioned on one another.""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """If true, the SQuAD examples contain some that do not have an answer."""} ) __SCREAMING_SNAKE_CASE = field( default=0.0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) __SCREAMING_SNAKE_CASE = field( default=20 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) __SCREAMING_SNAKE_CASE = field( default=0 , metadata={ """help""": ( """language id of input for language-specific xlm models (see""" """ tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)""" ) } , ) __SCREAMING_SNAKE_CASE = field(default=1 , metadata={"""help""": """multiple threads for converting example to features"""} ) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """train""" __SCREAMING_SNAKE_CASE = """dev""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 def __init__( self : str , a_ : SquadDataTrainingArguments , a_ : PreTrainedTokenizer , a_ : Optional[int] = None , a_ : Union[str, Split] = Split.train , a_ : Optional[bool] = False , a_ : Optional[str] = None , a_ : Optional[str] = "pt" , ): """simple docstring""" __snake_case = args __snake_case = is_language_sensitive __snake_case = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(a_ , a_ ): try: __snake_case = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) __snake_case = mode # Load data features from cache or dataset file __snake_case = "v2" if args.version_2_with_negative else "v1" __snake_case = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f'''cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}''' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __snake_case = cached_features_file + ".lock" with FileLock(a_ ): if os.path.exists(a_ ) and not args.overwrite_cache: __snake_case = time.time() __snake_case = torch.load(a_ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. __snake_case = self.old_features["features"] __snake_case = self.old_features.get("dataset" , a_ ) __snake_case = self.old_features.get("examples" , a_ ) logger.info( f'''Loading features from cached file {cached_features_file} [took %.3f s]''' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f'''Deleting cached file {cached_features_file} will allow dataset and examples to be cached in''' " future run" ) else: if mode == Split.dev: __snake_case = self.processor.get_dev_examples(args.data_dir ) else: __snake_case = self.processor.get_train_examples(args.data_dir ) __snake_case , __snake_case = squad_convert_examples_to_features( examples=self.examples , tokenizer=a_ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=a_ , ) __snake_case = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , a_ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f'''Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]''' ) def __len__( self : Optional[Any] ): """simple docstring""" return len(self.features ) def __getitem__( self : List[str] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.features[i] __snake_case = torch.tensor(feature.input_ids , dtype=torch.long ) __snake_case = torch.tensor(feature.attention_mask , dtype=torch.long ) __snake_case = torch.tensor(feature.token_type_ids , dtype=torch.long ) __snake_case = torch.tensor(feature.cls_index , dtype=torch.long ) __snake_case = torch.tensor(feature.p_mask , dtype=torch.float ) __snake_case = torch.tensor(feature.is_impossible , dtype=torch.float ) __snake_case = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: __snake_case = torch.tensor(feature.start_position , dtype=torch.long ) __snake_case = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = (KDPMaDiscreteScheduler,) __SCREAMING_SNAKE_CASE = 10 def A ( self : Any , **a_ : List[Any] ): """simple docstring""" __snake_case = { "num_train_timesteps": 1_100, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**a_ ) return config def A ( self : List[str] ): """simple docstring""" for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=a_ ) def A ( self : List[Any] ): """simple docstring""" for beta_start, beta_end in zip([0.00001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=a_ , beta_end=a_ ) def A ( self : int ): """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=a_ ) def A ( self : Union[str, Any] ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config(prediction_type="v_prediction" ) __snake_case = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps ) __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma __snake_case = sample.to(a_ ) for i, t in enumerate(scheduler.timesteps ): __snake_case = scheduler.scale_model_input(a_ , a_ ) __snake_case = model(a_ , a_ ) __snake_case = scheduler.step(a_ , a_ , a_ ) __snake_case = output.prev_sample __snake_case = torch.sum(torch.abs(a_ ) ) __snake_case = torch.mean(torch.abs(a_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6_934e-07 ) < 1e-2 assert abs(result_mean.item() - 6.1_112e-10 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 4.693_428_650_170_972e-07 ) < 1e-2 assert abs(result_mean.item() - 0.0002 ) < 1e-3 def A ( self : Any ): """simple docstring""" if torch_device == "mps": return __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps ) __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma __snake_case = sample.to(a_ ) for i, t in enumerate(scheduler.timesteps ): __snake_case = scheduler.scale_model_input(a_ , a_ ) __snake_case = model(a_ , a_ ) __snake_case = scheduler.step(a_ , a_ , a_ ) __snake_case = output.prev_sample __snake_case = torch.sum(torch.abs(a_ ) ) __snake_case = torch.mean(torch.abs(a_ ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3 def A ( self : Optional[Any] ): """simple docstring""" if torch_device == "mps": return __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps , device=a_ ) __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter.to(a_ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: __snake_case = scheduler.scale_model_input(a_ , a_ ) __snake_case = model(a_ , a_ ) __snake_case = scheduler.step(a_ , a_ , a_ ) __snake_case = output.prev_sample __snake_case = torch.sum(torch.abs(a_ ) ) __snake_case = torch.mean(torch.abs(a_ ) ) if str(a_ ).startswith("cpu" ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 20.4125 ) < 1e-2 assert abs(result_mean.item() - 0.0266 ) < 1e-3
69
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
1
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ): """simple docstring""" try: __snake_case = tempfile.mktemp() with open(a_ , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , a_ ) __snake_case = AlbertTokenizer.from_pretrained(a_ ) finally: os.remove(a_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , a_ ) __snake_case = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A ( self : str ): """simple docstring""" __snake_case = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : List[Any] ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a_ , repo_id="test-tokenizer" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a_ , repo_id="valid_org/test-tokenizer-org" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A ( self : List[str] ): """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizerFast.from_pretrained(a_ ) bert_tokenizer.save_pretrained(a_ ) __snake_case = CustomTokenizerFast.from_pretrained(a_ ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a_ , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A ( self : str ): """simple docstring""" __snake_case = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def A ( self : List[Any] ): """simple docstring""" __snake_case = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : str ): """simple docstring""" __snake_case = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def A ( self : Tuple ): """simple docstring""" __snake_case = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def A ( self : Any ): """simple docstring""" __snake_case = Trie() __snake_case = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a_ , ["AB", "C"] )
69
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Tuple = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """autoformer""" __SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : List[Any] , a_ : Optional[int] = None , a_ : Optional[int] = None , a_ : str = "student_t" , a_ : str = "nll" , a_ : int = 1 , a_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , a_ : bool = True , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : Optional[List[int]] = None , a_ : Optional[List[int]] = None , a_ : int = 64 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 32 , a_ : int = 32 , a_ : str = "gelu" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : int = 100 , a_ : float = 0.02 , a_ : bool = True , a_ : Union[str, Any]=True , a_ : int = 10 , a_ : int = 25 , a_ : int = 3 , **a_ : Tuple , ): """simple docstring""" __snake_case = prediction_length __snake_case = context_length if context_length is not None else prediction_length __snake_case = distribution_output __snake_case = loss __snake_case = input_size __snake_case = num_time_features __snake_case = lags_sequence __snake_case = scaling __snake_case = num_dynamic_real_features __snake_case = num_static_real_features __snake_case = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case = cardinality else: __snake_case = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case = embedding_dimension else: __snake_case = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case = num_parallel_samples # Transformer architecture configuration __snake_case = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case = d_model __snake_case = encoder_attention_heads __snake_case = decoder_attention_heads __snake_case = encoder_ffn_dim __snake_case = decoder_ffn_dim __snake_case = encoder_layers __snake_case = decoder_layers __snake_case = dropout __snake_case = attention_dropout __snake_case = activation_dropout __snake_case = encoder_layerdrop __snake_case = decoder_layerdrop __snake_case = activation_function __snake_case = init_std __snake_case = use_cache # Autoformer __snake_case = label_length __snake_case = moving_average __snake_case = autocorrelation_factor super().__init__(is_encoder_decoder=a_ , **a_ ) @property def A ( self : Optional[int] ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
69
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 a : Tuple = get_tests_dir('''fixtures''') a : Dict = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : int = get_tests_dir('''fixtures/dummy-config.json''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Tuple ): """simple docstring""" __snake_case = 0 def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __snake_case = AutoFeatureExtractor.from_pretrained(a_ ).to_dict() config_dict.pop("feature_extractor_type" ) __snake_case = WavaVecaFeatureExtractor(**a_ ) # save in new folder model_config.save_pretrained(a_ ) config.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) # make sure private variable is not incorrectly saved __snake_case = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with self.assertRaisesRegex( a_ , "bert-base is not a local folder and is not a valid model identifier" ): __snake_case = AutoFeatureExtractor.from_pretrained("bert-base" ) def A ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( a_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __snake_case = AutoFeatureExtractor.from_pretrained(a_ , revision="aaaaaa" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( a_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): __snake_case = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ , trust_remote_code=a_ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def A ( self : int ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoFeatureExtractor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def A ( self : Dict ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # If remote code is not set, the default is to use local __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(a_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
69
1
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import json import os import unittest from transformers import AutoTokenizer, GPTaTokenizer, GPTaTokenizerFast from transformers.models.gpta.tokenization_gpta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTaTokenizer __SCREAMING_SNAKE_CASE = GPTaTokenizerFast __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = {"""add_prefix_space""": True} __SCREAMING_SNAKE_CASE = False def A ( self : List[str] ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __snake_case = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", "<|endoftext|>", ] __snake_case = dict(zip(a_ , range(len(a_ ) ) ) ) __snake_case = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] __snake_case = {"unk_token": "<unk>"} __snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(a_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(a_ ) ) def A ( self : List[str] , **a_ : Union[str, Any] ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTaTokenizer.from_pretrained(self.tmpdirname , **a_ ) def A ( self : Union[str, Any] , **a_ : Tuple ): """simple docstring""" kwargs.update(self.special_tokens_map ) return GPTaTokenizerFast.from_pretrained(self.tmpdirname , **a_ ) def A ( self : List[str] , a_ : Any ): """simple docstring""" __snake_case = "lower newer" __snake_case = "lower newer" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = GPTaTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) __snake_case = "lower newer" __snake_case = ["\u0120low", "er", "\u0120", "n", "e", "w", "er"] __snake_case = tokenizer.tokenize(a_ , add_prefix_space=a_ ) self.assertListEqual(a_ , a_ ) __snake_case = tokens + [tokenizer.unk_token] __snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , a_ ) def A ( self : Any ): """simple docstring""" if not self.test_rust_tokenizer: return __snake_case = self.get_tokenizer() __snake_case = self.get_rust_tokenizer(add_prefix_space=a_ ) __snake_case = "lower newer" # Testing tokenization __snake_case = tokenizer.tokenize(a_ , add_prefix_space=a_ ) __snake_case = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_ , a_ ) # Testing conversion to ids without special tokens __snake_case = tokenizer.encode(a_ , add_special_tokens=a_ , add_prefix_space=a_ ) __snake_case = rust_tokenizer.encode(a_ , add_special_tokens=a_ ) self.assertListEqual(a_ , a_ ) # Testing conversion to ids with special tokens __snake_case = self.get_rust_tokenizer(add_prefix_space=a_ ) __snake_case = tokenizer.encode(a_ , add_prefix_space=a_ ) __snake_case = rust_tokenizer.encode(a_ ) self.assertListEqual(a_ , a_ ) # Testing the unknown token __snake_case = tokens + [rust_tokenizer.unk_token] __snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(a_ ) , a_ ) def A ( self : Any , *a_ : int , **a_ : Optional[Any] ): """simple docstring""" pass def A ( self : Union[str, Any] , a_ : Tuple=15 ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __snake_case = self.rust_tokenizer_class.from_pretrained(a_ , **a_ ) # Simple input __snake_case = "This is a simple input" __snake_case = ["This is a simple input 1", "This is a simple input 2"] __snake_case = ("This is a simple input", "This is a pair") __snake_case = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests self.assertRaises(a_ , tokenizer_r.encode , a_ , max_length=a_ , padding="max_length" ) # Simple input self.assertRaises(a_ , tokenizer_r.encode_plus , a_ , max_length=a_ , padding="max_length" ) # Simple input self.assertRaises( a_ , tokenizer_r.batch_encode_plus , a_ , max_length=a_ , padding="max_length" , ) # Pair input self.assertRaises(a_ , tokenizer_r.encode , a_ , max_length=a_ , padding="max_length" ) # Pair input self.assertRaises(a_ , tokenizer_r.encode_plus , a_ , max_length=a_ , padding="max_length" ) # Pair input self.assertRaises( a_ , tokenizer_r.batch_encode_plus , a_ , max_length=a_ , padding="max_length" , ) def A ( self : Any ): """simple docstring""" __snake_case = GPTaTokenizer.from_pretrained(self.tmpdirname , pad_token="<pad>" ) # Simple input __snake_case = "This is a simple input" __snake_case = ["This is a simple input looooooooong", "This is a simple input"] __snake_case = ("This is a simple input", "This is a pair") __snake_case = [ ("This is a simple input loooooong", "This is a simple input"), ("This is a simple pair loooooong", "This is a simple pair"), ] __snake_case = tokenizer.pad_token_id __snake_case = tokenizer(a_ , padding="max_length" , max_length=30 , return_tensors="np" ) __snake_case = tokenizer(a_ , padding=a_ , truncate=a_ , return_tensors="np" ) __snake_case = tokenizer(*a_ , padding="max_length" , max_length=60 , return_tensors="np" ) __snake_case = tokenizer(a_ , padding=a_ , truncate=a_ , return_tensors="np" ) # s # test single string max_length padding self.assertEqual(out_s["input_ids"].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s["input_ids"] ) self.assertTrue(0 in out_s["attention_mask"] ) # s2 # test automatic padding self.assertEqual(out_sa["input_ids"].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa["input_ids"][0] ) self.assertFalse(0 in out_sa["attention_mask"][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa["input_ids"][1] ) self.assertTrue(0 in out_sa["attention_mask"][1] ) # p # test single pair max_length padding self.assertEqual(out_p["input_ids"].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p["input_ids"] ) self.assertTrue(0 in out_p["attention_mask"] ) # p2 # test automatic padding pair self.assertEqual(out_pa["input_ids"].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa["input_ids"][0] ) self.assertFalse(0 in out_pa["attention_mask"][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa["input_ids"][1] ) self.assertTrue(0 in out_pa["attention_mask"][1] ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "$$$" __snake_case = GPTaTokenizer.from_pretrained(self.tmpdirname , bos_token=a_ , add_bos_token=a_ ) __snake_case = "This is a simple input" __snake_case = ["This is a simple input 1", "This is a simple input 2"] __snake_case = tokenizer.bos_token_id __snake_case = tokenizer(a_ ) __snake_case = tokenizer(a_ ) self.assertEqual(out_s.input_ids[0] , a_ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) __snake_case = tokenizer.decode(out_s.input_ids ) __snake_case = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , a_ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) def A ( self : Union[str, Any] ): """simple docstring""" pass def A ( self : Tuple ): """simple docstring""" __snake_case = [self.get_tokenizer(do_lower_case=a_ , add_bos_token=a_ )] for tokenizer in tokenizers: with self.subTest(f'''{tokenizer.__class__.__name__}''' ): __snake_case = "Encode this." __snake_case = "This one too please." __snake_case = tokenizer.encode(a_ , add_special_tokens=a_ ) encoded_sequence += tokenizer.encode(a_ , add_special_tokens=a_ ) __snake_case = tokenizer.encode_plus( a_ , a_ , add_special_tokens=a_ , return_special_tokens_mask=a_ , ) __snake_case = encoded_sequence_dict["input_ids"] __snake_case = encoded_sequence_dict["special_tokens_mask"] self.assertEqual(len(a_ ) , len(a_ ) ) __snake_case = [ (x if not special_tokens_mask[i] else None) for i, x in enumerate(a_ ) ] __snake_case = [x for x in filtered_sequence if x is not None] self.assertEqual(a_ , a_ ) @require_tokenizers class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("facebook/opt-350m" , from_slow=a_ ) __snake_case = "A photo of a cat" __snake_case = tokenizer.encode( a_ , ) self.assertEqual(a_ , [2, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained("test_opt" ) __snake_case = AutoTokenizer.from_pretrained("./test_opt" ) __snake_case = tokenizer.encode( a_ , ) self.assertEqual(a_ , [2, 250, 1_345, 9, 10, 4_758] ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("facebook/opt-350m" , use_slow=a_ ) __snake_case = "A photo of a cat" __snake_case = tokenizer.encode( a_ , ) # Same as above self.assertEqual(a_ , [2, 250, 1_345, 9, 10, 4_758] ) @unittest.skip("This test is failing because of a bug in the fast tokenizer" ) def A ( self : List[str] ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("facebook/opt-350m" , from_slow=a_ ) __snake_case = "bos" __snake_case = tokenizer.get_vocab()["bos"] __snake_case = "A photo of a cat" __snake_case = tokenizer.encode( a_ , ) # We changed the bos token self.assertEqual(a_ , [31_957, 250, 1_345, 9, 10, 4_758] ) tokenizer.save_pretrained("./tok" ) __snake_case = AutoTokenizer.from_pretrained("./tok" ) self.assertTrue(tokenizer.is_fast ) __snake_case = tokenizer.encode( a_ , ) self.assertEqual(a_ , [31_957, 250, 1_345, 9, 10, 4_758] )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a : Union[str, Any] = {'''configuration_xglm''': ['''XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XGLMConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : str = ['''XGLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = ['''XGLMTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = [ '''XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XGLMForCausalLM''', '''XGLMModel''', '''XGLMPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Union[str, Any] = [ '''FlaxXGLMForCausalLM''', '''FlaxXGLMModel''', '''FlaxXGLMPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Union[str, Any] = [ '''TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXGLMForCausalLM''', '''TFXGLMModel''', '''TFXGLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys a : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
69
'''simple docstring''' a : Dict = range(2, 20 + 1) a : Optional[int] = [10**k for k in range(ks[-1] + 1)] a : dict[int, dict[int, list[list[int]]]] = {} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] ) -> int: __snake_case = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) __snake_case = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) __snake_case , __snake_case = 0, 0 __snake_case = n - i __snake_case = memo.get(_UpperCAmelCase ) if sub_memo is not None: __snake_case = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over __snake_case = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __snake_case = _k break if max_jump >= 0: __snake_case , __snake_case , __snake_case = jumps[max_jump] # since the difference between jumps is cached, add c __snake_case = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: __snake_case = [] else: __snake_case = {c: []} __snake_case = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __snake_case , __snake_case = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __snake_case , __snake_case = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped __snake_case = sub_memo[c] # keep jumps sorted by # of terms skipped __snake_case = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __snake_case = i __snake_case , __snake_case , __snake_case = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __snake_case = ds_c + ds_b diff += addend __snake_case = 0 for j in range(_UpperCAmelCase ): __snake_case = a_i[j] + addend __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Tuple: for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case = digits[j] + addend if s >= 10: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) __snake_case = addend // 10 + quotient else: __snake_case = s __snake_case = addend // 10 if addend == 0: break while addend > 0: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int = 10**15 ) -> int: __snake_case = [1] __snake_case = 1 __snake_case = 0 while True: __snake_case , __snake_case = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break __snake_case = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
69
1
'''simple docstring''' import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """""" __SCREAMING_SNAKE_CASE = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) __SCREAMING_SNAKE_CASE = None # compression type in fsspec. ex: "gzip" __SCREAMING_SNAKE_CASE = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self : Tuple , a_ : str = "" , a_ : Optional[str] = None , a_ : Optional[dict] = None , **a_ : List[Any] ): """simple docstring""" super().__init__(self , **a_ ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode __snake_case = fsspec.open( a_ , mode="rb" , protocol=a_ , compression=self.compression , client_kwargs={ "requote_redirect_url": False, # see https://github.com/huggingface/datasets/pull/5459 "trust_env": True, # Enable reading proxy env variables. **(target_options or {}).pop("client_kwargs" , {} ), # To avoid issues if it was already passed. } , **(target_options or {}) , ) __snake_case = os.path.basename(self.file.path.split("::" )[0] ) __snake_case = ( self.compressed_name[: self.compressed_name.rindex("." )] if "." in self.compressed_name else self.compressed_name ) __snake_case = None @classmethod def A ( cls : Optional[int] , a_ : int ): """simple docstring""" return super()._strip_protocol(a_ ).lstrip("/" ) def A ( self : int ): """simple docstring""" if self.dir_cache is None: __snake_case = {**self.file.fs.info(self.file.path ), "name": self.uncompressed_name} __snake_case = {f["name"]: f} def A ( self : List[str] , a_ : str ): """simple docstring""" return self.file.open().read() def A ( self : Union[str, Any] , a_ : str , a_ : str = "rb" , a_ : List[str]=None , a_ : str=True , a_ : List[Any]=None , **a_ : int , ): """simple docstring""" __snake_case = self._strip_protocol(a_ ) if mode != "rb": raise ValueError(f'''Tried to read with mode {mode} on file {self.file.path} opened with mode \'rb\'''' ) return self.file.open() class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """bz2""" __SCREAMING_SNAKE_CASE = """bz2""" __SCREAMING_SNAKE_CASE = """.bz2""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """gzip""" __SCREAMING_SNAKE_CASE = """gzip""" __SCREAMING_SNAKE_CASE = """.gz""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """lz4""" __SCREAMING_SNAKE_CASE = """lz4""" __SCREAMING_SNAKE_CASE = """.lz4""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """xz""" __SCREAMING_SNAKE_CASE = """xz""" __SCREAMING_SNAKE_CASE = """.xz""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """zstd""" __SCREAMING_SNAKE_CASE = """zstd""" __SCREAMING_SNAKE_CASE = """.zst""" def __init__( self : Optional[int] , a_ : str , a_ : str = "rb" , a_ : Optional[str] = None , a_ : Optional[dict] = None , a_ : int = DEFAULT_BLOCK_SIZE , **a_ : int , ): """simple docstring""" super().__init__( fo=a_ , mode=a_ , target_protocol=a_ , target_options=a_ , block_size=a_ , **a_ , ) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 __snake_case = self.file.__enter__ class SCREAMING_SNAKE_CASE__ : def __init__( self : Any , a_ : Optional[int] ): """simple docstring""" __snake_case = file_ def __enter__( self : Dict ): """simple docstring""" self._file.__enter__() return self def __exit__( self : str , *a_ : Dict , **a_ : List[Any] ): """simple docstring""" self._file.__exit__(*a_ , **a_ ) def __iter__( self : List[Any] ): """simple docstring""" return iter(self._file ) def A ( self : Optional[int] ): """simple docstring""" return next(self._file ) def __getattr__( self : List[Any] , a_ : int ): """simple docstring""" return getattr(self._file , a_ ) def fixed_enter(*a_ : List[str] , **a_ : List[str] ): return WrappedFile(_enter(*a_ , **a_ ) ) __snake_case = fixed_enter
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
1
'''simple docstring''' from __future__ import annotations import math def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : bool , _UpperCAmelCase : list[int] , _UpperCAmelCase : float ) -> int: if depth < 0: raise ValueError("Depth cannot be less than 0" ) if not scores: raise ValueError("Scores cannot be empty" ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) , ) ) def __UpperCAmelCase ( ) -> None: __snake_case = [90, 23, 6, 33, 21, 65, 1_23, 3_44_23] __snake_case = math.log(len(_UpperCAmelCase ) , 2 ) print(F'''Optimal value : {minimax(0 , 0 , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase )}''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
69
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : list[int] , _UpperCAmelCase : int ) -> bool: __snake_case = len(_UpperCAmelCase ) __snake_case = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr value, a sum of zero(0) can be formed by not taking any element # hence True/1 for i in range(arr_len + 1 ): __snake_case = True # sum is not zero and set is empty then false for i in range(1 , required_sum + 1 ): __snake_case = False for i in range(1 , arr_len + 1 ): for j in range(1 , required_sum + 1 ): if arr[i - 1] > j: __snake_case = subset[i - 1][j] if arr[i - 1] <= j: __snake_case = subset[i - 1][j] or subset[i - 1][j - arr[i - 1]] return subset[arr_len][required_sum] if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import mpmath # for roots of unity import numpy as np class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , a_ : Optional[int]=None , a_ : int=None ): """simple docstring""" __snake_case = list(poly_a or [0] )[:] __snake_case = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() __snake_case = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() __snake_case = len(self.polyB ) # Add 0 to make lengths equal a power of 2 __snake_case = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform __snake_case = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product __snake_case = self.__multiply() def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" __snake_case = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB] # Corner case if len(a_ ) <= 1: return dft[0] # __snake_case = self.c_max_length // 2 while next_ncol > 0: __snake_case = [[] for i in range(a_ )] __snake_case = self.root**next_ncol # First half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update __snake_case = new_dft __snake_case = next_ncol // 2 return dft[0] def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.__dft("A" ) __snake_case = self.__dft("B" ) __snake_case = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT __snake_case = 2 while next_ncol <= self.c_max_length: __snake_case = [[] for i in range(a_ )] __snake_case = self.root ** (next_ncol // 2) __snake_case = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update __snake_case = new_inverse_c next_ncol *= 2 # Unpack __snake_case = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Optional[int] ): """simple docstring""" __snake_case = "A = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) __snake_case = "B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) __snake_case = "A*B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : Optional[Any] ) -> str: __snake_case = OmegaConf.load(_UpperCAmelCase ) __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] __snake_case = list(state_dict.keys() ) # extract state_dict for VQVAE __snake_case = {} __snake_case = "first_stage_model." for key in keys: if key.startswith(_UpperCAmelCase ): __snake_case = state_dict[key] # extract state_dict for UNetLDM __snake_case = {} __snake_case = "model.diffusion_model." for key in keys: if key.startswith(_UpperCAmelCase ): __snake_case = state_dict[key] __snake_case = config.model.params.first_stage_config.params __snake_case = config.model.params.unet_config.params __snake_case = VQModel(**_UpperCAmelCase ).eval() vqvae.load_state_dict(_UpperCAmelCase ) __snake_case = UNetLDMModel(**_UpperCAmelCase ).eval() unet.load_state_dict(_UpperCAmelCase ) __snake_case = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule="scaled_linear" , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=_UpperCAmelCase , ) __snake_case = LDMPipeline(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) pipeline.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() parser.add_argument('''--checkpoint_path''', type=str, required=True) parser.add_argument('''--config_path''', type=str, required=True) parser.add_argument('''--output_path''', type=str, required=True) a : Optional[int] = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
69
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
1
'''simple docstring''' from __future__ import annotations from collections.abc import MutableSequence class SCREAMING_SNAKE_CASE__ : def __init__( self : Optional[Any] , a_ : int , a_ : MutableSequence[float] ): """simple docstring""" if len(a_ ) != degree + 1: raise ValueError( "The number of coefficients should be equal to the degree + 1." ) __snake_case = list(a_ ) __snake_case = degree def __add__( self : List[str] , a_ : Polynomial ): """simple docstring""" if self.degree > polynomial_a.degree: __snake_case = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , a_ ) else: __snake_case = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , a_ ) def __sub__( self : Any , a_ : Polynomial ): """simple docstring""" return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self : Union[str, Any] ): """simple docstring""" return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self : Optional[int] , a_ : Polynomial ): """simple docstring""" __snake_case = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , a_ ) def A ( self : str , a_ : int | float ): """simple docstring""" __snake_case = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self : str ): """simple docstring""" __snake_case = "" for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(a_ ) return polynomial def __repr__( self : str ): """simple docstring""" return self.__str__() def A ( self : Tuple ): """simple docstring""" __snake_case = [0] * self.degree for i in range(self.degree ): __snake_case = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , a_ ) def A ( self : List[Any] , a_ : int | float = 0 ): """simple docstring""" __snake_case = [0] * (self.degree + 2) __snake_case = constant for i in range(self.degree + 1 ): __snake_case = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , a_ ) def __eq__( self : Union[str, Any] , a_ : object ): """simple docstring""" if not isinstance(a_ , a_ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self : int , a_ : object ): """simple docstring""" return not self.__eq__(a_ )
69
'''simple docstring''' import json import os import torch from diffusers import UNetaDModel os.makedirs('''hub/hopper-medium-v2/unet/hor32''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/unet/hor128''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/value_function''', exist_ok=True) def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> str: if hor == 1_28: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D") elif hor == 32: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 64, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D", "UpResnetBlock1D") __snake_case = torch.load(F'''/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch''' ) __snake_case = model.state_dict() __snake_case = { "down_block_types": down_block_types, "block_out_channels": block_out_channels, "up_block_types": up_block_types, "layers_per_block": 1, "use_timestep_embedding": True, "out_block_type": "OutConv1DBlock", "norm_num_groups": 8, "downsample_each_block": False, "in_channels": 14, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "flip_sin_to_cos": False, "freq_shift": 1, "sample_size": 6_55_36, "mid_block_type": "MidResTemporalBlock1D", "act_fn": "mish", } __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , F'''hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin''' ) with open(F'''hub/hopper-medium-v2/unet/hor{hor}/config.json''' , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) def __UpperCAmelCase ( ) -> List[Any]: __snake_case = { "in_channels": 14, "down_block_types": ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D"), "up_block_types": (), "out_block_type": "ValueFunction", "mid_block_type": "ValueFunctionMidBlock1D", "block_out_channels": (32, 64, 1_28, 2_56), "layers_per_block": 1, "downsample_each_block": True, "sample_size": 6_55_36, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "use_timestep_embedding": True, "flip_sin_to_cos": False, "freq_shift": 1, "norm_num_groups": 8, "act_fn": "mish", } __snake_case = torch.load("/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch" ) __snake_case = model __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , "hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin" ) with open("hub/hopper-medium-v2/value_function/config.json" , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": unet(32) # unet(128) value_function()
69
1
'''simple docstring''' from __future__ import annotations from PIL import Image # Define glider example a : List[str] = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example a : Tuple = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def __UpperCAmelCase ( _UpperCAmelCase : list[list[int]] ) -> list[list[int]]: __snake_case = [] for i in range(len(_UpperCAmelCase ) ): __snake_case = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours __snake_case = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(_UpperCAmelCase ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(_UpperCAmelCase ) - 1: neighbour_count += cells[i + 1][j] if i < len(_UpperCAmelCase ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. __snake_case = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(_UpperCAmelCase ) return next_generation def __UpperCAmelCase ( _UpperCAmelCase : list[list[int]] , _UpperCAmelCase : int ) -> list[Image.Image]: __snake_case = [] for _ in range(_UpperCAmelCase ): # Create output image __snake_case = Image.new("RGB" , (len(cells[0] ), len(_UpperCAmelCase )) ) __snake_case = img.load() # Save cells to image for x in range(len(_UpperCAmelCase ) ): for y in range(len(cells[0] ) ): __snake_case = 2_55 - cells[y][x] * 2_55 __snake_case = (colour, colour, colour) # Save image images.append(_UpperCAmelCase ) __snake_case = new_generation(_UpperCAmelCase ) return images if __name__ == "__main__": a : Optional[Any] = generate_images(GLIDER, 16) images[0].save('''out.gif''', save_all=True, append_images=images[1:])
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 1_00_00_00 ) -> int: __snake_case = 1 __snake_case = 1 __snake_case = {1: 1} for inputa in range(2 , _UpperCAmelCase ): __snake_case = 0 __snake_case = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __snake_case = (3 * number) + 1 counter += 1 if inputa not in counters: __snake_case = counter if counter > pre_counter: __snake_case = inputa __snake_case = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
69
1
'''simple docstring''' from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker a : List[Any] = '''CompVis/stable-diffusion-v1-1''' a : Optional[Any] = '''CompVis/stable-diffusion-v1-2''' a : Any = '''CompVis/stable-diffusion-v1-3''' a : Optional[int] = '''CompVis/stable-diffusion-v1-4''' class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Optional[int] , a_ : AutoencoderKL , a_ : CLIPTextModel , a_ : CLIPTokenizer , a_ : UNetaDConditionModel , a_ : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , a_ : StableDiffusionSafetyChecker , a_ : CLIPImageProcessor , a_ : bool = True , ): """simple docstring""" super()._init_() __snake_case = StableDiffusionPipeline.from_pretrained(a_ ) __snake_case = StableDiffusionPipeline.from_pretrained(a_ ) __snake_case = StableDiffusionPipeline.from_pretrained(a_ ) __snake_case = StableDiffusionPipeline( vae=a_ , text_encoder=a_ , tokenizer=a_ , unet=a_ , scheduler=a_ , safety_checker=a_ , feature_extractor=a_ , requires_safety_checker=a_ , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def A ( self : int ): """simple docstring""" return {k: getattr(self , a_ ) for k in self.config.keys() if not k.startswith("_" )} def A ( self : int , a_ : Optional[Union[str, int]] = "auto" ): """simple docstring""" if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(a_ ) def A ( self : str ): """simple docstring""" self.enable_attention_slicing(a_ ) @torch.no_grad() def A ( self : List[str] , a_ : Union[str, List[str]] , a_ : int = 512 , a_ : int = 512 , a_ : int = 50 , a_ : float = 7.5 , a_ : Optional[Union[str, List[str]]] = None , a_ : Optional[int] = 1 , a_ : float = 0.0 , a_ : Optional[torch.Generator] = None , a_ : Optional[torch.FloatTensor] = None , a_ : Optional[str] = "pil" , a_ : bool = True , a_ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a_ : int = 1 , **a_ : Optional[int] , ): """simple docstring""" return self.pipea( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) @torch.no_grad() def A ( self : int , a_ : Union[str, List[str]] , a_ : int = 512 , a_ : int = 512 , a_ : int = 50 , a_ : float = 7.5 , a_ : Optional[Union[str, List[str]]] = None , a_ : Optional[int] = 1 , a_ : float = 0.0 , a_ : Optional[torch.Generator] = None , a_ : Optional[torch.FloatTensor] = None , a_ : Optional[str] = "pil" , a_ : bool = True , a_ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a_ : int = 1 , **a_ : Union[str, Any] , ): """simple docstring""" return self.pipea( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) @torch.no_grad() def A ( self : Union[str, Any] , a_ : Union[str, List[str]] , a_ : int = 512 , a_ : int = 512 , a_ : int = 50 , a_ : float = 7.5 , a_ : Optional[Union[str, List[str]]] = None , a_ : Optional[int] = 1 , a_ : float = 0.0 , a_ : Optional[torch.Generator] = None , a_ : Optional[torch.FloatTensor] = None , a_ : Optional[str] = "pil" , a_ : bool = True , a_ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a_ : int = 1 , **a_ : Dict , ): """simple docstring""" return self.pipea( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) @torch.no_grad() def A ( self : Any , a_ : Union[str, List[str]] , a_ : int = 512 , a_ : int = 512 , a_ : int = 50 , a_ : float = 7.5 , a_ : Optional[Union[str, List[str]]] = None , a_ : Optional[int] = 1 , a_ : float = 0.0 , a_ : Optional[torch.Generator] = None , a_ : Optional[torch.FloatTensor] = None , a_ : Optional[str] = "pil" , a_ : bool = True , a_ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a_ : int = 1 , **a_ : Optional[Any] , ): """simple docstring""" return self.pipea( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) @torch.no_grad() def A ( self : int , a_ : Union[str, List[str]] , a_ : int = 512 , a_ : int = 512 , a_ : int = 50 , a_ : float = 7.5 , a_ : Optional[Union[str, List[str]]] = None , a_ : Optional[int] = 1 , a_ : float = 0.0 , a_ : Optional[torch.Generator] = None , a_ : Optional[torch.FloatTensor] = None , a_ : Optional[str] = "pil" , a_ : bool = True , a_ : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , a_ : int = 1 , **a_ : Union[str, Any] , ): """simple docstring""" __snake_case = "cuda" if torch.cuda.is_available() else "cpu" self.to(a_ ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` must be divisible by 8 but are {height} and {width}.''' ) # Get first result from Stable Diffusion Checkpoint v1.1 __snake_case = self.textaimg_sda_a( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) # Get first result from Stable Diffusion Checkpoint v1.2 __snake_case = self.textaimg_sda_a( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) # Get first result from Stable Diffusion Checkpoint v1.3 __snake_case = self.textaimg_sda_a( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) # Get first result from Stable Diffusion Checkpoint v1.4 __snake_case = self.textaimg_sda_a( prompt=a_ , height=a_ , width=a_ , num_inference_steps=a_ , guidance_scale=a_ , negative_prompt=a_ , num_images_per_prompt=a_ , eta=a_ , generator=a_ , latents=a_ , output_type=a_ , return_dict=a_ , callback=a_ , callback_steps=a_ , **a_ , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
69
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """SpeechT5FeatureExtractor""" __SCREAMING_SNAKE_CASE = """SpeechT5Tokenizer""" def __init__( self : List[Any] , a_ : str , a_ : str ): """simple docstring""" super().__init__(a_ , a_ ) def __call__( self : Dict , *a_ : Tuple , **a_ : List[str] ): """simple docstring""" __snake_case = kwargs.pop("audio" , a_ ) __snake_case = kwargs.pop("text" , a_ ) __snake_case = kwargs.pop("text_target" , a_ ) __snake_case = kwargs.pop("audio_target" , a_ ) __snake_case = kwargs.pop("sampling_rate" , a_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __snake_case = self.feature_extractor(a_ , *a_ , sampling_rate=a_ , **a_ ) elif text is not None: __snake_case = self.tokenizer(a_ , **a_ ) else: __snake_case = None if audio_target is not None: __snake_case = self.feature_extractor(audio_target=a_ , *a_ , sampling_rate=a_ , **a_ ) __snake_case = targets["input_values"] elif text_target is not None: __snake_case = self.tokenizer(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : str , **a_ : Dict ): """simple docstring""" __snake_case = kwargs.pop("input_values" , a_ ) __snake_case = kwargs.pop("input_ids" , a_ ) __snake_case = kwargs.pop("labels" , a_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) elif input_ids is not None: __snake_case = self.tokenizer.pad(a_ , **a_ ) else: __snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(a_ , a_ ) and "input_ids" in labels[0]): __snake_case = self.tokenizer.pad(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = self.feature_extractor.feature_size __snake_case = self.feature_extractor.num_mel_bins __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) __snake_case = feature_size_hack __snake_case = targets["input_values"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : Any , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : Optional[int] , *a_ : Union[str, Any] , **a_ : str ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ )
69
1
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Sequence, Value, load_dataset from transformers import DPRContextEncoder, DPRContextEncoderTokenizerFast, HfArgumentParser a : Any = logging.getLogger(__name__) torch.set_grad_enabled(False) a : Tuple = '''cuda''' if torch.cuda.is_available() else '''cpu''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : Tuple=1_00 , _UpperCAmelCase : List[Any]=" " ) -> List[str]: __snake_case = text.split(_UpperCAmelCase ) return [character.join(text[i : i + n] ).strip() for i in range(0 , len(_UpperCAmelCase ) , _UpperCAmelCase )] def __UpperCAmelCase ( _UpperCAmelCase : dict ) -> dict: __snake_case , __snake_case = [], [] for title, text in zip(documents["title"] , documents["text"] ): if text is not None: for passage in split_text(_UpperCAmelCase ): titles.append(title if title is not None else "" ) texts.append(_UpperCAmelCase ) return {"title": titles, "text": texts} def __UpperCAmelCase ( _UpperCAmelCase : dict , _UpperCAmelCase : DPRContextEncoder , _UpperCAmelCase : DPRContextEncoderTokenizerFast ) -> dict: __snake_case = ctx_tokenizer( documents["title"] , documents["text"] , truncation=_UpperCAmelCase , padding="longest" , return_tensors="pt" )["input_ids"] __snake_case = ctx_encoder(input_ids.to(device=_UpperCAmelCase ) , return_dict=_UpperCAmelCase ).pooler_output return {"embeddings": embeddings.detach().cpu().numpy()} def __UpperCAmelCase ( _UpperCAmelCase : "RagExampleArguments" , _UpperCAmelCase : "ProcessingArguments" , _UpperCAmelCase : "IndexHnswArguments" , ) -> Union[str, Any]: ###################################### logger.info("Step 1 - Create the dataset" ) ###################################### # The dataset needed for RAG must have three columns: # - title (string): title of the document # - text (string): text of a passage of the document # - embeddings (array of dimension d): DPR representation of the passage # Let's say you have documents in tab-separated csv files with columns "title" and "text" assert os.path.isfile(rag_example_args.csv_path ), "Please provide a valid path to a csv file" # You can load a Dataset object this way __snake_case = load_dataset( "csv" , data_files=[rag_example_args.csv_path] , split="train" , delimiter="\t" , column_names=["title", "text"] ) # More info about loading csv files in the documentation: https://huggingface.co/docs/datasets/loading_datasets.html?highlight=csv#csv-files # Then split the documents into passages of 100 words __snake_case = dataset.map(_UpperCAmelCase , batched=_UpperCAmelCase , num_proc=processing_args.num_proc ) # And compute the embeddings __snake_case = DPRContextEncoder.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ).to(device=_UpperCAmelCase ) __snake_case = DPRContextEncoderTokenizerFast.from_pretrained(rag_example_args.dpr_ctx_encoder_model_name ) __snake_case = Features( {"text": Value("string" ), "title": Value("string" ), "embeddings": Sequence(Value("float32" ) )} ) # optional, save as float32 instead of float64 to save space __snake_case = dataset.map( partial(_UpperCAmelCase , ctx_encoder=_UpperCAmelCase , ctx_tokenizer=_UpperCAmelCase ) , batched=_UpperCAmelCase , batch_size=processing_args.batch_size , features=_UpperCAmelCase , ) # And finally save your dataset __snake_case = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset" ) dataset.save_to_disk(_UpperCAmelCase ) # from datasets import load_from_disk # dataset = load_from_disk(passages_path) # to reload the dataset ###################################### logger.info("Step 2 - Index the dataset" ) ###################################### # Let's use the Faiss implementation of HNSW for fast approximate nearest neighbor search __snake_case = faiss.IndexHNSWFlat(index_hnsw_args.d , index_hnsw_args.m , faiss.METRIC_INNER_PRODUCT ) dataset.add_faiss_index("embeddings" , custom_index=_UpperCAmelCase ) # And save the index __snake_case = os.path.join(rag_example_args.output_dir , "my_knowledge_dataset_hnsw_index.faiss" ) dataset.get_index("embeddings" ).save(_UpperCAmelCase ) # dataset.load_faiss_index("embeddings", index_path) # to reload the index @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field( default=str(Path(_UpperCamelCase ).parent / """test_run""" / """dummy-kb""" / """my_knowledge_dataset.csv""" ) , metadata={"""help""": """Path to a tab-separated csv file with columns 'title' and 'text'"""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Question that is passed as input to RAG. Default is 'What does Moses' rod turn into ?'."""} , ) __SCREAMING_SNAKE_CASE = field( default="""facebook/rag-sequence-nq""" , metadata={"""help""": """The RAG model to use. Either 'facebook/rag-sequence-nq' or 'facebook/rag-token-nq'"""} , ) __SCREAMING_SNAKE_CASE = field( default="""facebook/dpr-ctx_encoder-multiset-base""" , metadata={ """help""": ( """The DPR context encoder model to use. Either 'facebook/dpr-ctx_encoder-single-nq-base' or""" """ 'facebook/dpr-ctx_encoder-multiset-base'""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=str(Path(_UpperCamelCase ).parent / """test_run""" / """dummy-kb""" ) , metadata={"""help""": """Path to a directory where the dataset passages and the index will be saved"""} , ) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": """The number of processes to use to split the documents into passages. Default is single process.""" } , ) __SCREAMING_SNAKE_CASE = field( default=16 , metadata={ """help""": """The batch size to use when computing the passages embeddings using the DPR context encoder.""" } , ) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field( default=768 , metadata={"""help""": """The dimension of the embeddings to pass to the HNSW Faiss index."""} , ) __SCREAMING_SNAKE_CASE = field( default=128 , metadata={ """help""": ( """The number of bi-directional links created for every new element during the HNSW index construction.""" ) } , ) if __name__ == "__main__": logging.basicConfig(level=logging.WARNING) logger.setLevel(logging.INFO) a : int = HfArgumentParser((RagExampleArguments, ProcessingArguments, IndexHnswArguments)) a , a , a : str = parser.parse_args_into_dataclasses() with TemporaryDirectory() as tmp_dir: a : str = rag_example_args.output_dir or tmp_dir main(rag_example_args, processing_args, index_hnsw_args)
69
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : Optional[Any] , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __snake_case = input_file.read() __snake_case = regexp.search(a_ ) return match def A ( self : Any , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __snake_case = regexp.finditer(a_ ) __snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A ( self : Optional[int] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a_ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(a_ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
1
'''simple docstring''' from typing import List, Optional, Union import numpy as np import torch import torchaudio.compliance.kaldi as ta_kaldi from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging a : Dict = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = ["""input_features""", """attention_mask"""] def __init__( self : Any , a_ : Optional[Any]=80 , a_ : Optional[Any]=16_000 , a_ : Tuple=80 , a_ : int=0.0 , a_ : List[Any]=True , a_ : Union[str, Any]=True , a_ : List[Any]=True , **a_ : List[Any] , ): """simple docstring""" super().__init__(feature_size=a_ , sampling_rate=a_ , padding_value=a_ , **a_ ) __snake_case = num_mel_bins __snake_case = do_ceptral_normalize __snake_case = normalize_means __snake_case = normalize_vars __snake_case = True def A ( self : Tuple , a_ : np.ndarray , ): """simple docstring""" __snake_case = waveform * (2**15) # Kaldi compliance: 16-bit signed integers __snake_case = torch.from_numpy(a_ ).unsqueeze(0 ) __snake_case = ta_kaldi.fbank(a_ , num_mel_bins=self.num_mel_bins , sample_frequency=self.sampling_rate ) return features.numpy() @staticmethod def A ( a_ : np.ndarray , a_ : int , a_ : Optional[bool] = True , a_ : Optional[bool] = True , a_ : float = 0.0 , ): """simple docstring""" if normalize_means: __snake_case = x[:input_length].mean(axis=0 ) __snake_case = np.subtract(a_ , a_ ) if normalize_vars: __snake_case = x[:input_length].std(axis=0 ) __snake_case = np.divide(a_ , a_ ) if input_length < x.shape[0]: __snake_case = padding_value # make sure array is in float32 __snake_case = x.astype(np.floataa ) return x def A ( self : Union[str, Any] , a_ : List[np.ndarray] , a_ : Optional[np.ndarray] = None ): """simple docstring""" __snake_case = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [ self.utterance_cmvn(a_ , a_ , self.normalize_means , self.normalize_vars , self.padding_value ) for x, n in zip(a_ , a_ ) ] def __call__( self : List[Any] , a_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a_ : Union[bool, str, PaddingStrategy] = False , a_ : Optional[int] = None , a_ : bool = False , a_ : Optional[int] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : Optional[int] = None , a_ : Optional[bool] = None , **a_ : Union[str, Any] , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) __snake_case = isinstance(a_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) __snake_case = is_batched_numpy or ( isinstance(a_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __snake_case = [np.asarray(a_ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a_ , np.ndarray ): __snake_case = np.asarray(a_ , dtype=np.floataa ) elif isinstance(a_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __snake_case = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __snake_case = [raw_speech] # extract fbank features __snake_case = [self._extract_fbank_features(a_ ) for waveform in raw_speech] # convert into correct format for padding __snake_case = BatchFeature({"input_features": features} ) __snake_case = self.pad( a_ , padding=a_ , max_length=a_ , truncation=a_ , pad_to_multiple_of=a_ , return_attention_mask=a_ , **a_ , ) # make sure list is in array format __snake_case = padded_inputs.get("input_features" ) if isinstance(input_features[0] , a_ ): __snake_case = [np.asarray(a_ , dtype=np.floataa ) for feature in input_features] __snake_case = padded_inputs.get("attention_mask" ) if attention_mask is not None: __snake_case = [np.asarray(a_ , dtype=np.intaa ) for array in attention_mask] # Utterance-level cepstral mean and variance normalization if self.do_ceptral_normalize: __snake_case = ( np.array(a_ , dtype=np.intaa ) if self._get_padding_strategies(a_ , max_length=a_ ) is not PaddingStrategy.DO_NOT_PAD else None ) __snake_case = self.normalize( padded_inputs["input_features"] , attention_mask=a_ ) if return_tensors is not None: __snake_case = padded_inputs.convert_to_tensors(a_ ) return padded_inputs
69
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : Dict = {'''vocab_file''': '''sentencepiece.model'''} a : Tuple = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } a : str = { '''google/rembert''': 256, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[Any] , a_ : int , a_ : Any=False , a_ : List[Any]=True , a_ : List[Any]=True , a_ : List[Any]="[CLS]" , a_ : List[Any]="[SEP]" , a_ : List[Any]="[UNK]" , a_ : str="[SEP]" , a_ : List[str]="[PAD]" , a_ : Optional[int]="[CLS]" , a_ : List[str]="[MASK]" , **a_ : str , ): """simple docstring""" super().__init__( do_lower_case=a_ , remove_space=a_ , keep_accents=a_ , bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , pad_token=a_ , cls_token=a_ , mask_token=a_ , **a_ , ) __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(a_ ) @property def A ( self : Optional[Any] ): """simple docstring""" return len(self.sp_model ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ): """simple docstring""" __snake_case = self.__dict__.copy() __snake_case = None return state def __setstate__( self : str , a_ : Optional[int] ): """simple docstring""" __snake_case = d __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def A ( self : Tuple , a_ : Optional[int] , a_ : int=False ): """simple docstring""" __snake_case = self.sp_model.EncodeAsPieces(a_ ) return pieces def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" return self.sp_model.PieceToId(a_ ) def A ( self : Optional[Any] , a_ : List[str] ): """simple docstring""" return self.sp_model.IdToPiece(a_ ) def A ( self : Optional[Any] , a_ : int ): """simple docstring""" __snake_case = self.sp_model.decode_pieces(a_ ) return out_string def A ( self : Union[str, Any] , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a_ )) + [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1] def A ( self : Tuple , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A ( self : List[Any] , a_ : str , a_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a_ ): logger.error("Vocabulary path ({}) should be a directory".format(a_ ) ) return __snake_case = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ): copyfile(self.vocab_file , a_ ) return (out_vocab_file,)
69
1
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable a : Union[str, Any] = { '''configuration_gpt_neox_japanese''': ['''GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoXJapaneseConfig'''], '''tokenization_gpt_neox_japanese''': ['''GPTNeoXJapaneseTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoXJapaneseForCausalLM''', '''GPTNeoXJapaneseLayer''', '''GPTNeoXJapaneseModel''', '''GPTNeoXJapanesePreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys a : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ): """simple docstring""" try: __snake_case = tempfile.mktemp() with open(a_ , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , a_ ) __snake_case = AlbertTokenizer.from_pretrained(a_ ) finally: os.remove(a_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , a_ ) __snake_case = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A ( self : str ): """simple docstring""" __snake_case = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : List[Any] ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a_ , repo_id="test-tokenizer" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a_ , repo_id="valid_org/test-tokenizer-org" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A ( self : List[str] ): """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizerFast.from_pretrained(a_ ) bert_tokenizer.save_pretrained(a_ ) __snake_case = CustomTokenizerFast.from_pretrained(a_ ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a_ , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A ( self : str ): """simple docstring""" __snake_case = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def A ( self : List[Any] ): """simple docstring""" __snake_case = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : str ): """simple docstring""" __snake_case = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def A ( self : Tuple ): """simple docstring""" __snake_case = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def A ( self : Any ): """simple docstring""" __snake_case = Trie() __snake_case = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a_ , ["AB", "C"] )
69
1
'''simple docstring''' import math def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: __snake_case = [True] * n __snake_case = False __snake_case = False __snake_case = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): __snake_case = i * 2 while index < n: __snake_case = False __snake_case = index + i __snake_case = [2] for i in range(3 , _UpperCAmelCase , 2 ): if is_prime[i]: primes.append(_UpperCAmelCase ) return primes def __UpperCAmelCase ( _UpperCAmelCase : int = 99_99_66_66_33_33 ) -> int: __snake_case = math.floor(math.sqrt(_UpperCAmelCase ) ) + 1_00 __snake_case = prime_sieve(_UpperCAmelCase ) __snake_case = 0 __snake_case = 0 __snake_case = primes[prime_index] while (last_prime**2) <= limit: __snake_case = primes[prime_index + 1] __snake_case = last_prime**2 __snake_case = next_prime**2 # Get numbers divisible by lps(current) __snake_case = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) __snake_case = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps __snake_case = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair __snake_case = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 __snake_case , __snake_case = 1, 1 for _ in range(number_of_steps - 1 ): __snake_case , __snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar a : str = TypeVar('''KEY''') a : Tuple = TypeVar('''VAL''') @dataclass(frozen=_UpperCamelCase , slots=_UpperCamelCase ) class SCREAMING_SNAKE_CASE__ ( Generic[KEY, VAL] ): __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 class SCREAMING_SNAKE_CASE__ ( _Item ): def __init__( self : Tuple ): """simple docstring""" super().__init__(a_ , a_ ) def __bool__( self : List[Any] ): """simple docstring""" return False a : Any = _DeletedItem() class SCREAMING_SNAKE_CASE__ ( MutableMapping[KEY, VAL] ): def __init__( self : Any , a_ : int = 8 , a_ : float = 0.75 ): """simple docstring""" __snake_case = initial_block_size __snake_case = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 __snake_case = capacity_factor __snake_case = 0 def A ( self : Union[str, Any] , a_ : KEY ): """simple docstring""" return hash(a_ ) % len(self._buckets ) def A ( self : Optional[Any] , a_ : int ): """simple docstring""" return (ind + 1) % len(self._buckets ) def A ( self : List[str] , a_ : int , a_ : KEY , a_ : VAL ): """simple docstring""" __snake_case = self._buckets[ind] if not stored: __snake_case = _Item(a_ , a_ ) self._len += 1 return True elif stored.key == key: __snake_case = _Item(a_ , a_ ) return True else: return False def A ( self : Tuple ): """simple docstring""" __snake_case = len(self._buckets ) * self._capacity_factor return len(self ) >= int(a_ ) def A ( self : Optional[int] ): """simple docstring""" if len(self._buckets ) <= self._initial_block_size: return False __snake_case = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def A ( self : str , a_ : int ): """simple docstring""" __snake_case = self._buckets __snake_case = [None] * new_size __snake_case = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def A ( self : List[Any] ): """simple docstring""" self._resize(len(self._buckets ) * 2 ) def A ( self : int ): """simple docstring""" self._resize(len(self._buckets ) // 2 ) def A ( self : List[Any] , a_ : KEY ): """simple docstring""" __snake_case = self._get_bucket_index(a_ ) for _ in range(len(self._buckets ) ): yield ind __snake_case = self._get_next_ind(a_ ) def A ( self : str , a_ : KEY , a_ : VAL ): """simple docstring""" for ind in self._iterate_buckets(a_ ): if self._try_set(a_ , a_ , a_ ): break def __setitem__( self : Union[str, Any] , a_ : KEY , a_ : VAL ): """simple docstring""" if self._is_full(): self._size_up() self._add_item(a_ , a_ ) def __delitem__( self : List[str] , a_ : KEY ): """simple docstring""" for ind in self._iterate_buckets(a_ ): __snake_case = self._buckets[ind] if item is None: raise KeyError(a_ ) if item is _deleted: continue if item.key == key: __snake_case = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self : Optional[int] , a_ : KEY ): """simple docstring""" for ind in self._iterate_buckets(a_ ): __snake_case = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(a_ ) def __len__( self : Dict ): """simple docstring""" return self._len def __iter__( self : List[str] ): """simple docstring""" yield from (item.key for item in self._buckets if item) def __repr__( self : int ): """simple docstring""" __snake_case = " ,".join( f'''{item.key}: {item.val}''' for item in self._buckets if item ) return f'''HashMap({val_string})'''
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str ) -> str: return " ".join( "".join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
69
1
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a : Optional[int] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = ["""pixel_values"""] def __init__( self : List[Any] , a_ : bool = True , a_ : Dict[str, int] = None , a_ : PILImageResampling = PILImageResampling.BICUBIC , a_ : bool = True , a_ : Union[int, float] = 1 / 255 , a_ : bool = True , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[float, List[float]]] = None , a_ : bool = True , **a_ : Optional[Any] , ): """simple docstring""" super().__init__(**a_ ) __snake_case = size if size is not None else {"height": 384, "width": 384} __snake_case = get_size_dict(a_ , default_to_square=a_ ) __snake_case = do_resize __snake_case = size __snake_case = resample __snake_case = do_rescale __snake_case = rescale_factor __snake_case = do_normalize __snake_case = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __snake_case = image_std if image_std is not None else OPENAI_CLIP_STD __snake_case = do_convert_rgb def A ( self : List[Any] , a_ : np.ndarray , a_ : Dict[str, int] , a_ : PILImageResampling = PILImageResampling.BICUBIC , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : str , ): """simple docstring""" __snake_case = get_size_dict(a_ , default_to_square=a_ ) if "height" not in size or "width" not in size: raise ValueError(f'''The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}''' ) __snake_case = (size["height"], size["width"]) return resize(a_ , size=a_ , resample=a_ , data_format=a_ , **a_ ) def A ( self : List[str] , a_ : np.ndarray , a_ : Union[int, float] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : int , ): """simple docstring""" return rescale(a_ , scale=a_ , data_format=a_ , **a_ ) def A ( self : Any , a_ : np.ndarray , a_ : Union[float, List[float]] , a_ : Union[float, List[float]] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Optional[Any] , ): """simple docstring""" return normalize(a_ , mean=a_ , std=a_ , data_format=a_ , **a_ ) def A ( self : Any , a_ : ImageInput , a_ : Optional[bool] = None , a_ : Optional[Dict[str, int]] = None , a_ : PILImageResampling = None , a_ : Optional[bool] = None , a_ : Optional[float] = None , a_ : Optional[bool] = None , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : bool = None , a_ : ChannelDimension = ChannelDimension.FIRST , **a_ : List[str] , ): """simple docstring""" __snake_case = do_resize if do_resize is not None else self.do_resize __snake_case = resample if resample is not None else self.resample __snake_case = do_rescale if do_rescale is not None else self.do_rescale __snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case = do_normalize if do_normalize is not None else self.do_normalize __snake_case = image_mean if image_mean is not None else self.image_mean __snake_case = image_std if image_std is not None else self.image_std __snake_case = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __snake_case = size if size is not None else self.size __snake_case = get_size_dict(a_ , default_to_square=a_ ) __snake_case = make_list_of_images(a_ ) if not valid_images(a_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: __snake_case = [convert_to_rgb(a_ ) for image in images] # All transformations expect numpy arrays. __snake_case = [to_numpy_array(a_ ) for image in images] if do_resize: __snake_case = [self.resize(image=a_ , size=a_ , resample=a_ ) for image in images] if do_rescale: __snake_case = [self.rescale(image=a_ , scale=a_ ) for image in images] if do_normalize: __snake_case = [self.normalize(image=a_ , mean=a_ , std=a_ ) for image in images] __snake_case = [to_channel_dimension_format(a_ , a_ ) for image in images] __snake_case = BatchFeature(data={"pixel_values": images} , tensor_type=a_ ) return encoded_outputs
69
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : Any , a_ : Union[str, Any]=13 , a_ : Any=7 , a_ : Any=True , a_ : Dict=True , a_ : Union[str, Any]=False , a_ : Tuple=True , a_ : str=99 , a_ : Tuple=64 , a_ : Tuple=5 , a_ : Union[str, Any]=4 , a_ : Dict=64 , a_ : Union[str, Any]="gelu" , a_ : Dict=0.1 , a_ : List[str]=0.1 , a_ : Dict=512 , a_ : Tuple=16 , a_ : str=2 , a_ : Any=0.02 , a_ : List[Any]=3 , a_ : Tuple=4 , a_ : Optional[int]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : int ): """simple docstring""" return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def A ( self : str ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def A ( self : Tuple , a_ : int , a_ : str , a_ : Optional[int] , a_ : List[Any] , a_ : str , a_ : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Any , a_ : int , a_ : Tuple , a_ : str , a_ : int , a_ : str , a_ : List[Any] ): """simple docstring""" __snake_case = MPNetForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Any , a_ : Any , a_ : int , a_ : Union[str, Any] , a_ : Dict , a_ : Optional[Any] , a_ : Any ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Optional[Any] , a_ : Any , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = MPNetForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict , a_ : List[str] , a_ : str , a_ : Union[str, Any] , a_ : str , a_ : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def A ( self : List[Any] ): """simple docstring""" __snake_case = MPNetModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel.from_pretrained("microsoft/mpnet-base" ) __snake_case = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a_ , atol=1e-4 ) )
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> int: if len(_UpperCAmelCase ) != len(_UpperCAmelCase ): raise ValueError("String lengths must match!" ) __snake_case = 0 for chara, chara in zip(_UpperCAmelCase , _UpperCAmelCase ): if chara != chara: count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Optional[int]: return 1 / (1 + np.exp(-z )) def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict ) -> List[str]: return (-y * np.log(_UpperCAmelCase ) - (1 - y) * np.log(1 - h )).mean() def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any] ) -> Optional[Any]: __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) return np.sum(y * scores - np.log(1 + np.exp(_UpperCAmelCase ) ) ) def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str]=7_00_00 ) -> Union[str, Any]: __snake_case = np.zeros(x.shape[1] ) for iterations in range(_UpperCAmelCase ): __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = np.dot(x.T , h - y ) / y.size __snake_case = theta - alpha * gradient # updating the weights __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = cost_function(_UpperCAmelCase , _UpperCAmelCase ) if iterations % 1_00 == 0: print(F'''loss: {j} \t''' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": a : int = datasets.load_iris() a : int = iris.data[:, :2] a : Optional[Any] = (iris.target != 0) * 1 a : Tuple = 0.1 a : List[str] = logistic_reg(alpha, x, y, max_iterations=70_000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> Union[str, Any]: return sigmoid_function( np.dot(_UpperCAmelCase , _UpperCAmelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((a) , (a)) : Any = (x[:, 0].min(), x[:, 0].max()) ((a) , (a)) : Any = (x[:, 1].min(), x[:, 1].max()) ((a) , (a)) : Any = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) a : Optional[Any] = np.c_[xxa.ravel(), xxa.ravel()] a : List[Any] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
69
1
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import ( AutoProcessor, BertTokenizerFast, BlipImageProcessor, GPTaTokenizer, InstructBlipProcessor, PreTrainedTokenizerFast, ) @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Dict ): """simple docstring""" __snake_case = tempfile.mkdtemp() __snake_case = BlipImageProcessor() __snake_case = GPTaTokenizer.from_pretrained("hf-internal-testing/tiny-random-GPT2Model" ) __snake_case = BertTokenizerFast.from_pretrained("hf-internal-testing/tiny-random-bert" ) __snake_case = InstructBlipProcessor(a_ , a_ , a_ ) processor.save_pretrained(self.tmpdirname ) def A ( self : Optional[Any] , **a_ : List[Any] ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **a_ ).tokenizer def A ( self : Any , **a_ : int ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **a_ ).image_processor def A ( self : Any , **a_ : List[str] ): """simple docstring""" return AutoProcessor.from_pretrained(self.tmpdirname , **a_ ).qformer_tokenizer def A ( self : List[Any] ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def A ( self : List[Any] ): """simple docstring""" __snake_case = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case = [Image.fromarray(np.moveaxis(a_ , 0 , -1 ) ) for x in image_inputs] return image_inputs def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = InstructBlipProcessor( tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() , qformer_tokenizer=self.get_qformer_tokenizer() , ) processor.save_pretrained(self.tmpdirname ) __snake_case = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) __snake_case = self.get_image_processor(do_normalize=a_ , padding_value=1.0 ) __snake_case = InstructBlipProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=a_ , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , a_ ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , a_ ) self.assertIsInstance(processor.qformer_tokenizer , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.get_image_processor() __snake_case = self.get_tokenizer() __snake_case = self.get_qformer_tokenizer() __snake_case = InstructBlipProcessor( tokenizer=a_ , image_processor=a_ , qformer_tokenizer=a_ ) __snake_case = self.prepare_image_inputs() __snake_case = image_processor(a_ , return_tensors="np" ) __snake_case = processor(images=a_ , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def A ( self : Any ): """simple docstring""" __snake_case = self.get_image_processor() __snake_case = self.get_tokenizer() __snake_case = self.get_qformer_tokenizer() __snake_case = InstructBlipProcessor( tokenizer=a_ , image_processor=a_ , qformer_tokenizer=a_ ) __snake_case = "lower newer" __snake_case = processor(text=a_ ) __snake_case = tokenizer(a_ , return_token_type_ids=a_ ) __snake_case = qformer_tokenizer(a_ , return_token_type_ids=a_ ) for key in encoded_tokens.keys(): self.assertListEqual(encoded_tokens[key] , encoded_processor[key] ) for key in encoded_tokens_qformer.keys(): self.assertListEqual(encoded_tokens_qformer[key] , encoded_processor["qformer_" + key] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.get_image_processor() __snake_case = self.get_tokenizer() __snake_case = self.get_qformer_tokenizer() __snake_case = InstructBlipProcessor( tokenizer=a_ , image_processor=a_ , qformer_tokenizer=a_ ) __snake_case = "lower newer" __snake_case = self.prepare_image_inputs() __snake_case = processor(text=a_ , images=a_ ) self.assertListEqual( list(inputs.keys() ) , ["input_ids", "attention_mask", "qformer_input_ids", "qformer_attention_mask", "pixel_values"] , ) # test if it raises when no input is passed with pytest.raises(a_ ): processor() def A ( self : Tuple ): """simple docstring""" __snake_case = self.get_image_processor() __snake_case = self.get_tokenizer() __snake_case = self.get_qformer_tokenizer() __snake_case = InstructBlipProcessor( tokenizer=a_ , image_processor=a_ , qformer_tokenizer=a_ ) __snake_case = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __snake_case = processor.batch_decode(a_ ) __snake_case = tokenizer.batch_decode(a_ ) self.assertListEqual(a_ , a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.get_image_processor() __snake_case = self.get_tokenizer() __snake_case = self.get_qformer_tokenizer() __snake_case = InstructBlipProcessor( tokenizer=a_ , image_processor=a_ , qformer_tokenizer=a_ ) __snake_case = "lower newer" __snake_case = self.prepare_image_inputs() __snake_case = processor(text=a_ , images=a_ ) self.assertListEqual( list(inputs.keys() ) , ["input_ids", "attention_mask", "qformer_input_ids", "qformer_attention_mask", "pixel_values"] , )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : Any , a_ : Union[str, Any]=13 , a_ : Any=7 , a_ : Any=True , a_ : Dict=True , a_ : Union[str, Any]=False , a_ : Tuple=True , a_ : str=99 , a_ : Tuple=64 , a_ : Tuple=5 , a_ : Union[str, Any]=4 , a_ : Dict=64 , a_ : Union[str, Any]="gelu" , a_ : Dict=0.1 , a_ : List[str]=0.1 , a_ : Dict=512 , a_ : Tuple=16 , a_ : str=2 , a_ : Any=0.02 , a_ : List[Any]=3 , a_ : Tuple=4 , a_ : Optional[int]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : int ): """simple docstring""" return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def A ( self : str ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def A ( self : Tuple , a_ : int , a_ : str , a_ : Optional[int] , a_ : List[Any] , a_ : str , a_ : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Any , a_ : int , a_ : Tuple , a_ : str , a_ : int , a_ : str , a_ : List[Any] ): """simple docstring""" __snake_case = MPNetForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Any , a_ : Any , a_ : int , a_ : Union[str, Any] , a_ : Dict , a_ : Optional[Any] , a_ : Any ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Optional[Any] , a_ : Any , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = MPNetForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict , a_ : List[str] , a_ : str , a_ : Union[str, Any] , a_ : str , a_ : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def A ( self : List[Any] ): """simple docstring""" __snake_case = MPNetModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel.from_pretrained("microsoft/mpnet-base" ) __snake_case = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a_ , atol=1e-4 ) )
69
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a : Optional[Any] = { '''configuration_graphormer''': ['''GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GraphormerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = [ '''GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GraphormerForGraphClassification''', '''GraphormerModel''', '''GraphormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys a : str = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Tuple = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """autoformer""" __SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : List[Any] , a_ : Optional[int] = None , a_ : Optional[int] = None , a_ : str = "student_t" , a_ : str = "nll" , a_ : int = 1 , a_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , a_ : bool = True , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : Optional[List[int]] = None , a_ : Optional[List[int]] = None , a_ : int = 64 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 32 , a_ : int = 32 , a_ : str = "gelu" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : int = 100 , a_ : float = 0.02 , a_ : bool = True , a_ : Union[str, Any]=True , a_ : int = 10 , a_ : int = 25 , a_ : int = 3 , **a_ : Tuple , ): """simple docstring""" __snake_case = prediction_length __snake_case = context_length if context_length is not None else prediction_length __snake_case = distribution_output __snake_case = loss __snake_case = input_size __snake_case = num_time_features __snake_case = lags_sequence __snake_case = scaling __snake_case = num_dynamic_real_features __snake_case = num_static_real_features __snake_case = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case = cardinality else: __snake_case = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case = embedding_dimension else: __snake_case = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case = num_parallel_samples # Transformer architecture configuration __snake_case = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case = d_model __snake_case = encoder_attention_heads __snake_case = decoder_attention_heads __snake_case = encoder_ffn_dim __snake_case = decoder_ffn_dim __snake_case = encoder_layers __snake_case = decoder_layers __snake_case = dropout __snake_case = attention_dropout __snake_case = activation_dropout __snake_case = encoder_layerdrop __snake_case = decoder_layerdrop __snake_case = activation_function __snake_case = init_std __snake_case = use_cache # Autoformer __snake_case = label_length __snake_case = moving_average __snake_case = autocorrelation_factor super().__init__(is_encoder_decoder=a_ , **a_ ) @property def A ( self : Optional[int] ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
69
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a : List[str] = {'''configuration_swin''': ['''SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SwinConfig''', '''SwinOnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[str] = [ '''SWIN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SwinForImageClassification''', '''SwinForMaskedImageModeling''', '''SwinModel''', '''SwinPreTrainedModel''', '''SwinBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = [ '''TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFSwinForImageClassification''', '''TFSwinForMaskedImageModeling''', '''TFSwinModel''', '''TFSwinPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys a : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
1
'''simple docstring''' from __future__ import annotations from typing import Any class SCREAMING_SNAKE_CASE__ : def __init__( self : Optional[int] , a_ : int , a_ : int , a_ : float = 0 ): """simple docstring""" __snake_case , __snake_case = row, column __snake_case = [[default_value for c in range(a_ )] for r in range(a_ )] def __str__( self : Dict ): """simple docstring""" __snake_case = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier __snake_case = 0 for row_vector in self.array: for obj in row_vector: __snake_case = max(a_ , len(str(a_ ) ) ) __snake_case = f'''%{max_element_length}s''' # Make string and return def single_line(a_ : list[float] ) -> str: nonlocal string_format_identifier __snake_case = "[" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(a_ ) for row_vector in self.array ) return s def __repr__( self : Any ): """simple docstring""" return str(self ) def A ( self : List[str] , a_ : tuple[int, int] ): """simple docstring""" if not (isinstance(a_ , (list, tuple) ) and len(a_ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple , a_ : tuple[int, int] ): """simple docstring""" assert self.validate_indicies(a_ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple , a_ : tuple[int, int] , a_ : float ): """simple docstring""" assert self.validate_indicies(a_ ) __snake_case = value def __add__( self : Any , a_ : Matrix ): """simple docstring""" assert isinstance(a_ , a_ ) assert self.row == another.row and self.column == another.column # Add __snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __snake_case = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): """simple docstring""" __snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __snake_case = -self[r, c] return result def __sub__( self : int , a_ : Matrix ): """simple docstring""" return self + (-another) def __mul__( self : Tuple , a_ : int | float | Matrix ): """simple docstring""" if isinstance(a_ , (int, float) ): # Scalar multiplication __snake_case = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __snake_case = self[r, c] * another return result elif isinstance(a_ , a_ ): # Matrix multiplication assert self.column == another.row __snake_case = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __snake_case = f'''Unsupported type given for another ({type(a_ )})''' raise TypeError(a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): __snake_case = self[r, c] return result def A ( self : List[Any] , a_ : Matrix , a_ : Matrix ): """simple docstring""" assert isinstance(a_ , a_ ) and isinstance(a_ , a_ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __snake_case = v.transpose() __snake_case = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def __UpperCAmelCase ( ) -> None: # a^(-1) __snake_case = Matrix(3 , 3 , 0 ) for i in range(3 ): __snake_case = 1 print(F'''a^(-1) is {ainv}''' ) # u, v __snake_case = Matrix(3 , 1 , 0 ) __snake_case , __snake_case , __snake_case = 1, 2, -3 __snake_case = Matrix(3 , 1 , 0 ) __snake_case , __snake_case , __snake_case = 4, -2, 5 print(F'''u is {u}''' ) print(F'''v is {v}''' ) print(F'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(_UpperCAmelCase , _UpperCAmelCase )}''' ) def __UpperCAmelCase ( ) -> None: import doctest doctest.testmod() testa()
69
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 a : Tuple = get_tests_dir('''fixtures''') a : Dict = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : int = get_tests_dir('''fixtures/dummy-config.json''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Tuple ): """simple docstring""" __snake_case = 0 def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __snake_case = AutoFeatureExtractor.from_pretrained(a_ ).to_dict() config_dict.pop("feature_extractor_type" ) __snake_case = WavaVecaFeatureExtractor(**a_ ) # save in new folder model_config.save_pretrained(a_ ) config.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) # make sure private variable is not incorrectly saved __snake_case = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with self.assertRaisesRegex( a_ , "bert-base is not a local folder and is not a valid model identifier" ): __snake_case = AutoFeatureExtractor.from_pretrained("bert-base" ) def A ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( a_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __snake_case = AutoFeatureExtractor.from_pretrained(a_ , revision="aaaaaa" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( a_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): __snake_case = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ , trust_remote_code=a_ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def A ( self : int ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoFeatureExtractor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def A ( self : Dict ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # If remote code is not set, the default is to use local __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(a_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
69
1
'''simple docstring''' import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): a : Dict = True from torch.cuda.amp import autocast a : str = logging.getLogger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=None , _UpperCAmelCase : Any=None ) -> Optional[int]: return field(default_factory=lambda: default , metadata=_UpperCAmelCase ) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) __SCREAMING_SNAKE_CASE = field( default=0.1 , metadata={"""help""": """The dropout ratio for the attention probabilities."""} ) __SCREAMING_SNAKE_CASE = field( default=0.1 , metadata={"""help""": """The dropout ratio for activations inside the fully connected layer."""} ) __SCREAMING_SNAKE_CASE = field( default=0.1 , metadata={ """help""": """The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.""" } , ) __SCREAMING_SNAKE_CASE = field( default=0.1 , metadata={"""help""": """The dropout probabilitiy for all 1D convolutional layers in feature extractor."""} , ) __SCREAMING_SNAKE_CASE = field( default=0.05 , metadata={ """help""": ( """Propability of each feature vector along the time axis to be chosen as the start of the vector""" """span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature""" """vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.""" ) } , ) __SCREAMING_SNAKE_CASE = field(default=0.0 , metadata={"""help""": """The LayerDrop probability."""} ) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) __SCREAMING_SNAKE_CASE = field( default="""train+validation""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to 'train'""" } , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of validation examples to this """ """value if set.""" ) } , ) __SCREAMING_SNAKE_CASE = list_field( default=[""",""", """?""", """.""", """!""", """-""", """;""", """:""", """\"\"""", """%""", """'""", """\"""", """�"""] , metadata={"""help""": """A list of characters to remove from the transcripts."""} , ) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None def __call__( self : Tuple , a_ : List[Dict[str, Union[List[int], torch.Tensor]]] ): """simple docstring""" __snake_case = [{"input_values": feature["input_values"]} for feature in features] __snake_case = [{"input_ids": feature["labels"]} for feature in features] __snake_case = self.processor.pad( a_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" , ) __snake_case = self.processor.pad( labels=a_ , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors="pt" , ) # replace padding with -100 to ignore loss correctly __snake_case = labels_batch["input_ids"].masked_fill(labels_batch.attention_mask.ne(1 ) , -100 ) __snake_case = labels return batch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : List[str] , a_ : nn.Module , a_ : Dict[str, Union[torch.Tensor, Any]] ): """simple docstring""" model.train() __snake_case = self._prepare_inputs(a_ ) if self.use_amp: with autocast(): __snake_case = self.compute_loss(a_ , a_ ) else: __snake_case = self.compute_loss(a_ , a_ ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": __snake_case = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": __snake_case = loss.sum() / (inputs["labels"] >= 0).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: __snake_case = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(a_ ).backward() elif self.use_apex: with amp.scale_loss(a_ , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(a_ ) else: loss.backward() return loss.detach() def __UpperCAmelCase ( ) -> Optional[Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. __snake_case , __snake_case , __snake_case = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: __snake_case , __snake_case , __snake_case = parser.parse_args_into_dataclasses() # Detecting last checkpoint. __snake_case = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: __snake_case = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info("Training/evaluation parameters %s" , _UpperCAmelCase ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: __snake_case = datasets.load_dataset( "common_voice" , data_args.dataset_config_name , split=data_args.train_split_name ) __snake_case = datasets.load_dataset("common_voice" , data_args.dataset_config_name , split="test" ) # Create and save tokenizer __snake_case = F'''[{"".join(data_args.chars_to_ignore )}]''' def remove_special_characters(_UpperCAmelCase : Dict ): __snake_case = re.sub(_UpperCAmelCase , "" , batch["sentence"] ).lower() + " " return batch __snake_case = train_dataset.map(_UpperCAmelCase , remove_columns=["sentence"] ) __snake_case = eval_dataset.map(_UpperCAmelCase , remove_columns=["sentence"] ) def extract_all_chars(_UpperCAmelCase : Tuple ): __snake_case = " ".join(batch["text"] ) __snake_case = list(set(_UpperCAmelCase ) ) return {"vocab": [vocab], "all_text": [all_text]} __snake_case = train_dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , batch_size=-1 , keep_in_memory=_UpperCAmelCase , remove_columns=train_dataset.column_names , ) __snake_case = train_dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , batch_size=-1 , keep_in_memory=_UpperCAmelCase , remove_columns=eval_dataset.column_names , ) __snake_case = list(set(vocab_train["vocab"][0] ) | set(vocab_test["vocab"][0] ) ) __snake_case = {v: k for k, v in enumerate(_UpperCAmelCase )} __snake_case = vocab_dict[" "] del vocab_dict[" "] __snake_case = len(_UpperCAmelCase ) __snake_case = len(_UpperCAmelCase ) with open("vocab.json" , "w" ) as vocab_file: json.dump(_UpperCAmelCase , _UpperCAmelCase ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __snake_case = WavaVecaCTCTokenizer( "vocab.json" , unk_token="[UNK]" , pad_token="[PAD]" , word_delimiter_token="|" , ) __snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0.0 , do_normalize=_UpperCAmelCase , return_attention_mask=_UpperCAmelCase ) __snake_case = WavaVecaProcessor(feature_extractor=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) __snake_case = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction="mean" , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: __snake_case = min(len(_UpperCAmelCase ) , data_args.max_train_samples ) __snake_case = train_dataset.select(range(_UpperCAmelCase ) ) if data_args.max_val_samples is not None: __snake_case = eval_dataset.select(range(data_args.max_val_samples ) ) __snake_case = torchaudio.transforms.Resample(4_80_00 , 1_60_00 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(_UpperCAmelCase : Tuple ): __snake_case , __snake_case = torchaudio.load(batch["path"] ) __snake_case = resampler(_UpperCAmelCase ).squeeze().numpy() __snake_case = 1_60_00 __snake_case = batch["text"] return batch __snake_case = train_dataset.map( _UpperCAmelCase , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) __snake_case = eval_dataset.map( _UpperCAmelCase , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(_UpperCAmelCase : Dict ): # check that all files have the correct sampling rate assert ( len(set(batch["sampling_rate"] ) ) == 1 ), F'''Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.''' __snake_case = processor( audio=batch["speech"] , text=batch["target_text"] , sampling_rate=batch["sampling_rate"][0] ) batch.update(_UpperCAmelCase ) return batch __snake_case = train_dataset.map( _UpperCAmelCase , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=_UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , ) __snake_case = eval_dataset.map( _UpperCAmelCase , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=_UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , ) # Metric __snake_case = datasets.load_metric("wer" ) def compute_metrics(_UpperCAmelCase : List[str] ): __snake_case = pred.predictions __snake_case = np.argmax(_UpperCAmelCase , axis=-1 ) __snake_case = processor.tokenizer.pad_token_id __snake_case = processor.batch_decode(_UpperCAmelCase ) # we do not want to group tokens when computing the metrics __snake_case = processor.batch_decode(pred.label_ids , group_tokens=_UpperCAmelCase ) __snake_case = wer_metric.compute(predictions=_UpperCAmelCase , references=_UpperCAmelCase ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator __snake_case = DataCollatorCTCWithPadding(processor=_UpperCAmelCase , padding=_UpperCAmelCase ) # Initialize our Trainer __snake_case = CTCTrainer( model=_UpperCAmelCase , data_collator=_UpperCAmelCase , args=_UpperCAmelCase , compute_metrics=_UpperCAmelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: __snake_case = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): __snake_case = model_args.model_name_or_path else: __snake_case = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) __snake_case = trainer.train(resume_from_checkpoint=_UpperCAmelCase ) trainer.save_model() __snake_case = train_result.metrics __snake_case = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_UpperCAmelCase ) ) __snake_case = min(_UpperCAmelCase , len(_UpperCAmelCase ) ) trainer.log_metrics("train" , _UpperCAmelCase ) trainer.save_metrics("train" , _UpperCAmelCase ) trainer.save_state() # Evaluation __snake_case = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) __snake_case = trainer.evaluate() __snake_case = data_args.max_val_samples if data_args.max_val_samples is not None else len(_UpperCAmelCase ) __snake_case = min(_UpperCAmelCase , len(_UpperCAmelCase ) ) trainer.log_metrics("eval" , _UpperCAmelCase ) trainer.save_metrics("eval" , _UpperCAmelCase ) return results if __name__ == "__main__": main()
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' from __future__ import annotations def __UpperCAmelCase ( _UpperCAmelCase : dict , _UpperCAmelCase : str ) -> set[str]: __snake_case , __snake_case = set(_UpperCAmelCase ), [start] while stack: __snake_case = stack.pop() explored.add(_UpperCAmelCase ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(_UpperCAmelCase ) return explored a : List[str] = { '''A''': ['''B''', '''C''', '''D'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F'''], '''D''': ['''B''', '''D'''], '''E''': ['''B''', '''F'''], '''F''': ['''C''', '''E''', '''G'''], '''G''': ['''F'''], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, '''A'''))
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline __SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS __SCREAMING_SNAKE_CASE = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __SCREAMING_SNAKE_CASE = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __SCREAMING_SNAKE_CASE = frozenset([] ) def A ( self : Tuple ): """simple docstring""" torch.manual_seed(0 ) __snake_case = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=a_ , ) __snake_case = PNDMScheduler(skip_prk_steps=a_ ) torch.manual_seed(0 ) __snake_case = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __snake_case = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="gelu" , projection_dim=512 , ) __snake_case = CLIPTextModel(a_ ) __snake_case = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) __snake_case = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def A ( self : Any , a_ : Optional[Any] , a_ : List[Any]=0 ): """simple docstring""" __snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(a_ ) ).to(a_ ) __snake_case = image.cpu().permute(0 , 2 , 3 , 1 )[0] __snake_case = Image.fromarray(np.uinta(a_ ) ).convert("RGB" ).resize((64, 64) ) __snake_case = Image.fromarray(np.uinta(image + 4 ) ).convert("RGB" ).resize((64, 64) ) if str(a_ ).startswith("mps" ): __snake_case = torch.manual_seed(a_ ) else: __snake_case = torch.Generator(device=a_ ).manual_seed(a_ ) __snake_case = { "prompt": "A painting of a squirrel eating a burger", "image": init_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def A ( self : Optional[Any] ): """simple docstring""" __snake_case = "cpu" # ensure determinism for the device-dependent torch.Generator __snake_case = self.get_dummy_components() __snake_case = StableDiffusionInpaintPipeline(**a_ ) __snake_case = sd_pipe.to(a_ ) sd_pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_dummy_inputs(a_ ) __snake_case = sd_pipe(**a_ ).images __snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __snake_case = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def A ( self : Dict ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Any ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : int ): """simple docstring""" __snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) __snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) __snake_case = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench.npy" ) __snake_case = "stabilityai/stable-diffusion-2-inpainting" __snake_case = StableDiffusionInpaintPipeline.from_pretrained(a_ , safety_checker=a_ ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) pipe.enable_attention_slicing() __snake_case = "Face of a yellow cat, high resolution, sitting on a park bench" __snake_case = torch.manual_seed(0 ) __snake_case = pipe( prompt=a_ , image=a_ , mask_image=a_ , generator=a_ , output_type="np" , ) __snake_case = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9e-3 def A ( self : int ): """simple docstring""" __snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) __snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) __snake_case = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench_fp16.npy" ) __snake_case = "stabilityai/stable-diffusion-2-inpainting" __snake_case = StableDiffusionInpaintPipeline.from_pretrained( a_ , torch_dtype=torch.floataa , safety_checker=a_ , ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) pipe.enable_attention_slicing() __snake_case = "Face of a yellow cat, high resolution, sitting on a park bench" __snake_case = torch.manual_seed(0 ) __snake_case = pipe( prompt=a_ , image=a_ , mask_image=a_ , generator=a_ , output_type="np" , ) __snake_case = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5e-1 def A ( self : List[Any] ): """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) __snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) __snake_case = "stabilityai/stable-diffusion-2-inpainting" __snake_case = PNDMScheduler.from_pretrained(a_ , subfolder="scheduler" ) __snake_case = StableDiffusionInpaintPipeline.from_pretrained( a_ , safety_checker=a_ , scheduler=a_ , torch_dtype=torch.floataa , ) pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __snake_case = "Face of a yellow cat, high resolution, sitting on a park bench" __snake_case = torch.manual_seed(0 ) __snake_case = pipe( prompt=a_ , image=a_ , mask_image=a_ , generator=a_ , num_inference_steps=2 , output_type="np" , ) __snake_case = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
69
'''simple docstring''' a : Dict = range(2, 20 + 1) a : Optional[int] = [10**k for k in range(ks[-1] + 1)] a : dict[int, dict[int, list[list[int]]]] = {} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] ) -> int: __snake_case = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) __snake_case = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) __snake_case , __snake_case = 0, 0 __snake_case = n - i __snake_case = memo.get(_UpperCAmelCase ) if sub_memo is not None: __snake_case = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over __snake_case = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __snake_case = _k break if max_jump >= 0: __snake_case , __snake_case , __snake_case = jumps[max_jump] # since the difference between jumps is cached, add c __snake_case = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: __snake_case = [] else: __snake_case = {c: []} __snake_case = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __snake_case , __snake_case = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __snake_case , __snake_case = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped __snake_case = sub_memo[c] # keep jumps sorted by # of terms skipped __snake_case = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __snake_case = i __snake_case , __snake_case , __snake_case = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __snake_case = ds_c + ds_b diff += addend __snake_case = 0 for j in range(_UpperCAmelCase ): __snake_case = a_i[j] + addend __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Tuple: for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case = digits[j] + addend if s >= 10: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) __snake_case = addend // 10 + quotient else: __snake_case = s __snake_case = addend // 10 if addend == 0: break while addend > 0: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int = 10**15 ) -> int: __snake_case = [1] __snake_case = 1 __snake_case = 0 while True: __snake_case , __snake_case = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break __snake_case = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
69
1
'''simple docstring''' import os import jsonlines import numpy as np from tqdm import tqdm a : int = 2_048 a : Optional[int] = 4_096 a : Dict = 42 a : Optional[int] = os.environ.pop('''PROCESS_TRAIN''', '''false''') a : List[str] = {'''null''': 0, '''short''': 1, '''long''': 2, '''yes''': 3, '''no''': 4} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: def choose_first(_UpperCAmelCase : str , _UpperCAmelCase : int=False ): assert isinstance(_UpperCAmelCase , _UpperCAmelCase ) if len(_UpperCAmelCase ) == 1: __snake_case = answer[0] return {k: [answer[k]] for k in answer} if is_long_answer else answer for a in answer: if is_long_answer: __snake_case = {k: [a[k]] for k in a} if len(a["start_token"] ) > 0: break return a __snake_case = {"id": example["id"]} __snake_case = example["annotations"] __snake_case = annotation["yes_no_answer"] if 0 in yes_no_answer or 1 in yes_no_answer: __snake_case = ["yes"] if 1 in yes_no_answer else ["no"] __snake_case = __snake_case = [] __snake_case = __snake_case = [] __snake_case = ["<cls>"] else: __snake_case = ["short"] __snake_case = choose_first(annotation["short_answers"] ) if len(out["start_token"] ) == 0: # answer will be long if short is not available __snake_case = ["long"] __snake_case = choose_first(annotation["long_answer"] , is_long_answer=_UpperCAmelCase ) __snake_case = [] answer.update(_UpperCAmelCase ) # disregard some samples if len(answer["start_token"] ) > 1 or answer["start_token"] == answer["end_token"]: __snake_case = True else: __snake_case = False __snake_case = ["start_token", "end_token", "start_byte", "end_byte", "text"] if not all(isinstance(answer[k] , _UpperCAmelCase ) for k in cols ): raise ValueError("Issue in ID" , example["id"] ) return answer def __UpperCAmelCase ( _UpperCAmelCase : Dict , _UpperCAmelCase : List[Any]=False ) -> List[Any]: __snake_case = _get_single_answer(_UpperCAmelCase ) # bytes are of no use del answer["start_byte"] del answer["end_byte"] # handle yes_no answers explicitly if answer["category"][0] in ["yes", "no"]: # category is list with one element __snake_case = example["document"]["tokens"] __snake_case = [] for i in range(len(doc["token"] ) ): if not doc["is_html"][i]: context.append(doc["token"][i] ) return { "context": " ".join(_UpperCAmelCase ), "answer": { "start_token": -1_00, # ignore index in cross-entropy "end_token": -1_00, # ignore index in cross-entropy "category": answer["category"], "span": answer["category"], # extra }, } # later, help in removing all no answers if answer["start_token"] == [-1]: return { "context": "None", "answer": { "start_token": -1, "end_token": -1, "category": "null", "span": "None", # extra }, } # handling normal samples __snake_case = ["start_token", "end_token"] answer.update({k: answer[k][0] if len(answer[k] ) > 0 else answer[k] for k in cols} ) # e.g. [10] == 10 __snake_case = example["document"]["tokens"] __snake_case = answer["start_token"] __snake_case = answer["end_token"] __snake_case = [] for i in range(len(doc["token"] ) ): if not doc["is_html"][i]: context.append(doc["token"][i] ) else: if answer["start_token"] > i: start_token -= 1 if answer["end_token"] > i: end_token -= 1 __snake_case = " ".join(context[start_token:end_token] ) # checking above code if assertion: __snake_case = doc["is_html"][answer["start_token"] : answer["end_token"]] __snake_case = doc["token"][answer["start_token"] : answer["end_token"]] __snake_case = " ".join([old[i] for i in range(len(_UpperCAmelCase ) ) if not is_html[i]] ) if new != old: print("ID:" , example["id"] ) print("New:" , _UpperCAmelCase , end="\n" ) print("Old:" , _UpperCAmelCase , end="\n\n" ) return { "context": " ".join(_UpperCAmelCase ), "answer": { "start_token": start_token, "end_token": end_token - 1, # this makes it inclusive "category": answer["category"], # either long or short "span": new, # extra }, } def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Any=20_48 , _UpperCAmelCase : str=40_96 , _UpperCAmelCase : Any=True ) -> Optional[Any]: # overlap will be of doc_stride - q_len __snake_case = get_context_and_ans(_UpperCAmelCase , assertion=_UpperCAmelCase ) __snake_case = out["answer"] # later, removing these samples if answer["start_token"] == -1: return { "example_id": example["id"], "input_ids": [[-1]], "labels": { "start_token": [-1], "end_token": [-1], "category": ["null"], }, } __snake_case = tokenizer(example["question"]["text"] , out["context"] ).input_ids __snake_case = input_ids.index(tokenizer.sep_token_id ) + 1 # return yes/no if answer["category"][0] in ["yes", "no"]: # category is list with one element __snake_case = [] __snake_case = [] __snake_case = input_ids[:q_len] __snake_case = range(_UpperCAmelCase , len(_UpperCAmelCase ) , max_length - doc_stride ) for i in doc_start_indices: __snake_case = i + max_length - q_len __snake_case = input_ids[i:end_index] inputs.append(q_indices + slice ) category.append(answer["category"][0] ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": [-1_00] * len(_UpperCAmelCase ), "end_token": [-1_00] * len(_UpperCAmelCase ), "category": category, }, } __snake_case = out["context"].split() __snake_case = splitted_context[answer["end_token"]] __snake_case = len( tokenizer( " ".join(splitted_context[: answer["start_token"]] ) , add_special_tokens=_UpperCAmelCase , ).input_ids ) __snake_case = len( tokenizer(" ".join(splitted_context[: answer["end_token"]] ) , add_special_tokens=_UpperCAmelCase ).input_ids ) answer["start_token"] += q_len answer["end_token"] += q_len # fixing end token __snake_case = len(tokenizer(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids ) if num_sub_tokens > 1: answer["end_token"] += num_sub_tokens - 1 __snake_case = input_ids[answer["start_token"] : answer["end_token"] + 1] # right & left are inclusive __snake_case = answer["start_token"] __snake_case = answer["end_token"] if assertion: __snake_case = tokenizer.decode(_UpperCAmelCase ) if answer["span"] != new: print("ISSUE IN TOKENIZATION" ) print("OLD:" , answer["span"] ) print("NEW:" , _UpperCAmelCase , end="\n\n" ) if len(_UpperCAmelCase ) <= max_length: return { "example_id": example["id"], "input_ids": [input_ids], "labels": { "start_token": [answer["start_token"]], "end_token": [answer["end_token"]], "category": answer["category"], }, } __snake_case = input_ids[:q_len] __snake_case = range(_UpperCAmelCase , len(_UpperCAmelCase ) , max_length - doc_stride ) __snake_case = [] __snake_case = [] __snake_case = [] __snake_case = [] # null, yes, no, long, short for i in doc_start_indices: __snake_case = i + max_length - q_len __snake_case = input_ids[i:end_index] inputs.append(q_indices + slice ) assert len(inputs[-1] ) <= max_length, "Issue in truncating length" if start_token >= i and end_token <= end_index - 1: __snake_case = start_token - i + q_len __snake_case = end_token - i + q_len answers_category.append(answer["category"][0] ) # ["short"] -> "short" else: __snake_case = -1_00 __snake_case = -1_00 answers_category.append("null" ) __snake_case = inputs[-1][start_token : end_token + 1] answers_start_token.append(_UpperCAmelCase ) answers_end_token.append(_UpperCAmelCase ) if assertion: if new != old and new != [tokenizer.cls_token_id]: print("ISSUE in strided for ID:" , example["id"] ) print("New:" , tokenizer.decode(_UpperCAmelCase ) ) print("Old:" , tokenizer.decode(_UpperCAmelCase ) , end="\n\n" ) if slice[-1] == tokenizer.sep_token_id: break return { "example_id": example["id"], "input_ids": inputs, "labels": { "start_token": answers_start_token, "end_token": answers_end_token, "category": answers_category, }, } def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : str=20_48 , _UpperCAmelCase : List[Any]=40_96 , _UpperCAmelCase : Optional[Any]=False ) -> Tuple: __snake_case = get_strided_contexts_and_ans( _UpperCAmelCase , _UpperCAmelCase , doc_stride=_UpperCAmelCase , max_length=_UpperCAmelCase , assertion=_UpperCAmelCase , ) return example def __UpperCAmelCase ( _UpperCAmelCase : Any , _UpperCAmelCase : Dict ) -> int: with jsonlines.open(_UpperCAmelCase , "a" ) as writer: for example in tqdm(_UpperCAmelCase , total=len(_UpperCAmelCase ) , desc="Saving samples ... " ): __snake_case = example["labels"] for ids, start, end, cat in zip( example["input_ids"] , labels["start_token"] , labels["end_token"] , labels["category"] , ): if start == -1 and end == -1: continue # leave waste samples with no answer if cat == "null" and np.random.rand() < 0.6: continue # removing 50 % samples writer.write( { "input_ids": ids, "start_token": start, "end_token": end, "category": CATEGORY_MAPPING[cat], } ) if __name__ == "__main__": from datasets import load_dataset from transformers import BigBirdTokenizer a : Dict = load_dataset('''natural_questions''') a : List[str] = BigBirdTokenizer.from_pretrained('''google/bigbird-roberta-base''') a : Optional[Any] = data['''train''' if PROCESS_TRAIN == '''true''' else '''validation'''] a : Union[str, Any] = { '''tokenizer''': tokenizer, '''doc_stride''': DOC_STRIDE, '''max_length''': MAX_LENGTH, '''assertion''': False, } a : Tuple = data.map(prepare_inputs, fn_kwargs=fn_kwargs) a : int = data.remove_columns(['''annotations''', '''document''', '''id''', '''question''']) print(data) np.random.seed(SEED) a : Union[str, Any] = '''nq-training.jsonl''' if PROCESS_TRAIN == '''true''' else '''nq-validation.jsonl''' save_to_disk(data, file_name=cache_file_name)
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : List[Any] , a_ : Any , a_ : Union[str, Any]=13 , a_ : Dict=7 , a_ : int=True , a_ : Union[str, Any]=True , a_ : Union[str, Any]=True , a_ : List[Any]=True , a_ : Optional[Any]=99 , a_ : str=32 , a_ : List[str]=2 , a_ : List[Any]=4 , a_ : Optional[Any]=37 , a_ : int="gelu" , a_ : Dict=0.1 , a_ : Union[str, Any]=0.1 , a_ : List[str]=512 , a_ : List[Any]=16 , a_ : Union[str, Any]=2 , a_ : Any=0.02 , a_ : Tuple=3 , a_ : Tuple=4 , a_ : int=None , ): """simple docstring""" __snake_case = parent __snake_case = 13 __snake_case = 7 __snake_case = True __snake_case = True __snake_case = True __snake_case = True __snake_case = 99 __snake_case = 32 __snake_case = 2 __snake_case = 4 __snake_case = 37 __snake_case = "gelu" __snake_case = 0.1 __snake_case = 0.1 __snake_case = 512 __snake_case = 16 __snake_case = 2 __snake_case = 0.02 __snake_case = 3 __snake_case = 4 __snake_case = None def A ( self : List[str] ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None if self.use_token_type_ids: __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=a_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Union[str, Any] , a_ : List[Any] , a_ : List[str] , a_ : Dict , a_ : str , a_ : Dict , a_ : List[str] , a_ : List[Any] ): """simple docstring""" __snake_case = TFRoFormerModel(config=a_ ) __snake_case = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} __snake_case = [input_ids, input_mask] __snake_case = model(a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[str] , a_ : str , a_ : int , a_ : Any , a_ : Tuple , a_ : List[str] , a_ : Any , a_ : List[Any] ): """simple docstring""" __snake_case = True __snake_case = TFRoFormerForCausalLM(config=a_ ) __snake_case = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __snake_case = model(a_ )["logits"] self.parent.assertListEqual( list(prediction_scores.numpy().shape ) , [self.batch_size, self.seq_length, self.vocab_size] ) def A ( self : Any , a_ : List[Any] , a_ : Tuple , a_ : List[Any] , a_ : Any , a_ : List[Any] , a_ : str , a_ : List[str] ): """simple docstring""" __snake_case = TFRoFormerForMaskedLM(config=a_ ) __snake_case = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __snake_case = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : str , a_ : Tuple , a_ : Tuple , a_ : str , a_ : Tuple , a_ : Dict , a_ : Any , a_ : Dict ): """simple docstring""" __snake_case = self.num_labels __snake_case = TFRoFormerForSequenceClassification(config=a_ ) __snake_case = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __snake_case = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : int , a_ : int , a_ : str , a_ : Any , a_ : Optional[Any] , a_ : Any , a_ : Tuple , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = TFRoFormerForMultipleChoice(config=a_ ) __snake_case = tf.tile(tf.expand_dims(a_ , 1 ) , (1, self.num_choices, 1) ) __snake_case = tf.tile(tf.expand_dims(a_ , 1 ) , (1, self.num_choices, 1) ) __snake_case = tf.tile(tf.expand_dims(a_ , 1 ) , (1, self.num_choices, 1) ) __snake_case = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } __snake_case = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Optional[int] , a_ : Any , a_ : List[Any] , a_ : List[Any] , a_ : Any , a_ : List[str] , a_ : Dict , a_ : Tuple ): """simple docstring""" __snake_case = self.num_labels __snake_case = TFRoFormerForTokenClassification(config=a_ ) __snake_case = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __snake_case = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : Dict , a_ : str , a_ : int , a_ : Any , a_ : List[str] , a_ : Optional[int] , a_ : List[Any] , a_ : str ): """simple docstring""" __snake_case = TFRoFormerForQuestionAnswering(config=a_ ) __snake_case = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } __snake_case = model(a_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Dict ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": TFRoFormerModel, """fill-mask""": TFRoFormerForMaskedLM, """question-answering""": TFRoFormerForQuestionAnswering, """text-classification""": TFRoFormerForSequenceClassification, """text-generation""": TFRoFormerForCausalLM, """token-classification""": TFRoFormerForTokenClassification, """zero-shot""": TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Any , a_ : Union[str, Any] , a_ : List[str] , a_ : Any , a_ : List[Any] , a_ : Dict ): """simple docstring""" if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def A ( self : Tuple ): """simple docstring""" __snake_case = TFRoFormerModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*a_ ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*a_ ) def A ( self : Any ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a_ ) @slow def A ( self : Dict ): """simple docstring""" __snake_case = TFRoFormerModel.from_pretrained("junnyu/roformer_chinese_base" ) self.assertIsNotNone(a_ ) @require_tf class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : List[Any] ): """simple docstring""" __snake_case = TFRoFormerForMaskedLM.from_pretrained("junnyu/roformer_chinese_base" ) __snake_case = tf.constant([[0, 1, 2, 3, 4, 5]] ) __snake_case = model(a_ )[0] # TODO Replace vocab size __snake_case = 50_000 __snake_case = [1, 6, vocab_size] self.assertEqual(output.shape , a_ ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. __snake_case = tf.constant( [ [ [-0.12053341, -1.0264901, 0.29221946], [-1.5133783, 0.197433, 0.15190607], [-5.0135403, -3.900256, -0.84038764], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , a_ , atol=1e-4 ) @require_tf class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = 1E-4 def A ( self : Optional[int] ): """simple docstring""" __snake_case = tf.constant([[4, 10]] ) __snake_case = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6 , embedding_dim=6 ) __snake_case = emba(input_ids.shape ) __snake_case = tf.constant( [[0.0000, 0.0000, 0.0000, 1.0000, 1.0000, 1.0000], [0.8415, 0.0464, 0.0022, 0.5403, 0.9989, 1.0000]] ) tf.debugging.assert_near(a_ , a_ , atol=self.tolerance ) def A ( self : List[str] ): """simple docstring""" __snake_case = tf.constant( [ [0.0000, 0.0000, 0.0000, 0.0000, 0.0000], [0.8415, 0.8219, 0.8020, 0.7819, 0.7617], [0.9093, 0.9364, 0.9581, 0.9749, 0.9870], ] ) __snake_case = TFRoFormerSinusoidalPositionalEmbedding(num_positions=512 , embedding_dim=512 ) emba([2, 16, 512] ) __snake_case = emba.weight[:3, :5] tf.debugging.assert_near(a_ , a_ , atol=self.tolerance ) @require_tf class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = 1E-4 def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 __snake_case = -tf.reshape(tf.range(2 * 12 * 16 * 64 , dtype=tf.floataa ) , shape=(2, 12, 16, 64) ) / 100 __snake_case = TFRoFormerSinusoidalPositionalEmbedding(num_positions=32 , embedding_dim=64 ) __snake_case = embed_positions([2, 16, 768] )[None, None, :, :] __snake_case , __snake_case = TFRoFormerSelfAttention.apply_rotary_position_embeddings( a_ , a_ , a_ ) __snake_case = tf.constant( [ [0.0000, 0.0100, 0.0200, 0.0300, 0.0400, 0.0500, 0.0600, 0.0700], [-0.2012, 0.8897, 0.0263, 0.9401, 0.2074, 0.9463, 0.3481, 0.9343], [-1.7057, 0.6271, -1.2145, 1.3897, -0.6303, 1.7647, -0.1173, 1.8985], [-2.1731, -1.6397, -2.7358, 0.2854, -2.1840, 1.7183, -1.3018, 2.4871], [0.2717, -3.6173, -2.9206, -2.1988, -3.6638, 0.3858, -2.9155, 2.2980], [3.9859, -2.1580, -0.7984, -4.4904, -4.1181, -2.0252, -4.4782, 1.1253], ] ) __snake_case = tf.constant( [ [0.0000, -0.0100, -0.0200, -0.0300, -0.0400, -0.0500, -0.0600, -0.0700], [0.2012, -0.8897, -0.0263, -0.9401, -0.2074, -0.9463, -0.3481, -0.9343], [1.7057, -0.6271, 1.2145, -1.3897, 0.6303, -1.7647, 0.1173, -1.8985], [2.1731, 1.6397, 2.7358, -0.2854, 2.1840, -1.7183, 1.3018, -2.4871], [-0.2717, 3.6173, 2.9206, 2.1988, 3.6638, -0.3858, 2.9155, -2.2980], [-3.9859, 2.1580, 0.7984, 4.4904, 4.1181, 2.0252, 4.4782, -1.1253], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8] , a_ , atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8] , a_ , atol=self.tolerance )
69
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
1
'''simple docstring''' from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer a : int = logging.get_logger(__name__) a : List[Any] = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } a : Dict = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } a : List[Any] = { '''facebook/blenderbot_small-90M''': 512, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = BlenderbotSmallTokenizer def __init__( self : str , a_ : int=None , a_ : str=None , a_ : str="<|endoftext|>" , a_ : Any="<|endoftext|>" , a_ : str="<|endoftext|>" , a_ : Any=False , a_ : Union[str, Any]=True , **a_ : str , ): """simple docstring""" super().__init__( ByteLevelBPETokenizer( vocab=a_ , merges=a_ , add_prefix_space=a_ , trim_offsets=a_ , ) , bos_token=a_ , eos_token=a_ , unk_token=a_ , **a_ , ) __snake_case = add_prefix_space def A ( self : Tuple , a_ : int , a_ : Any=None ): """simple docstring""" __snake_case = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
69
'''simple docstring''' import mpmath # for roots of unity import numpy as np class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , a_ : Optional[int]=None , a_ : int=None ): """simple docstring""" __snake_case = list(poly_a or [0] )[:] __snake_case = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() __snake_case = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() __snake_case = len(self.polyB ) # Add 0 to make lengths equal a power of 2 __snake_case = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform __snake_case = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product __snake_case = self.__multiply() def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" __snake_case = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB] # Corner case if len(a_ ) <= 1: return dft[0] # __snake_case = self.c_max_length // 2 while next_ncol > 0: __snake_case = [[] for i in range(a_ )] __snake_case = self.root**next_ncol # First half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update __snake_case = new_dft __snake_case = next_ncol // 2 return dft[0] def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.__dft("A" ) __snake_case = self.__dft("B" ) __snake_case = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT __snake_case = 2 while next_ncol <= self.c_max_length: __snake_case = [[] for i in range(a_ )] __snake_case = self.root ** (next_ncol // 2) __snake_case = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update __snake_case = new_inverse_c next_ncol *= 2 # Unpack __snake_case = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Optional[int] ): """simple docstring""" __snake_case = "A = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) __snake_case = "B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) __snake_case = "A*B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import unittest from transformers import SqueezeBertConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, ) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Optional[Any] , a_ : List[Any] , a_ : Tuple=13 , a_ : Optional[Any]=7 , a_ : List[Any]=True , a_ : List[str]=True , a_ : List[str]=False , a_ : List[str]=True , a_ : Union[str, Any]=99 , a_ : List[Any]=32 , a_ : Optional[Any]=5 , a_ : Union[str, Any]=4 , a_ : Optional[int]=64 , a_ : Any="gelu" , a_ : Optional[int]=0.1 , a_ : List[str]=0.1 , a_ : str=512 , a_ : int=16 , a_ : Optional[int]=2 , a_ : Dict=0.02 , a_ : List[Any]=3 , a_ : Any=4 , a_ : Tuple=None , a_ : Optional[int]=2 , a_ : int=2 , a_ : Union[str, Any]=2 , a_ : Union[str, Any]=2 , a_ : List[Any]=4 , a_ : List[Any]=1 , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope __snake_case = q_groups __snake_case = k_groups __snake_case = v_groups __snake_case = post_attention_groups __snake_case = intermediate_groups __snake_case = output_groups def A ( self : Dict ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : str ): """simple docstring""" return SqueezeBertConfig( embedding_size=self.hidden_size , vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , attention_probs_dropout_prob=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , q_groups=self.q_groups , k_groups=self.k_groups , v_groups=self.v_groups , post_attention_groups=self.post_attention_groups , intermediate_groups=self.intermediate_groups , output_groups=self.output_groups , ) def A ( self : Dict , a_ : Optional[Any] , a_ : List[Any] , a_ : List[str] , a_ : List[Any] , a_ : Dict , a_ : Dict ): """simple docstring""" __snake_case = SqueezeBertModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Dict , a_ : Optional[Any] , a_ : int , a_ : Any , a_ : int , a_ : Tuple , a_ : Dict ): """simple docstring""" __snake_case = SqueezeBertForMaskedLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Any , a_ : List[Any] , a_ : str , a_ : Any , a_ : List[str] , a_ : Tuple , a_ : str ): """simple docstring""" __snake_case = SqueezeBertForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Optional[Any] , a_ : Tuple , a_ : Tuple , a_ : Dict , a_ : List[Any] , a_ : Dict , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = SqueezeBertForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Tuple , a_ : List[str] , a_ : Optional[Any] , a_ : Optional[Any] , a_ : Optional[int] , a_ : List[Any] , a_ : Optional[int] ): """simple docstring""" __snake_case = self.num_labels __snake_case = SqueezeBertForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : int , a_ : str , a_ : int , a_ : Union[str, Any] , a_ : Optional[int] , a_ : Dict , a_ : Dict ): """simple docstring""" __snake_case = self.num_choices __snake_case = SqueezeBertForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : int ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( SqueezeBertModel, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, ) if is_torch_available() else None ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": SqueezeBertModel, """fill-mask""": SqueezeBertForMaskedLM, """question-answering""": SqueezeBertForQuestionAnswering, """text-classification""": SqueezeBertForSequenceClassification, """token-classification""": SqueezeBertForTokenClassification, """zero-shot""": SqueezeBertForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = SqueezeBertModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , dim=37 ) def A ( self : Dict ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_model(*a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_masked_lm(*a_ ) def A ( self : str ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_question_answering(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_token_classification(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_squeezebert_for_multiple_choice(*a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for model_name in SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = SqueezeBertModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) @require_sentencepiece @require_tokenizers @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = SqueezeBertForSequenceClassification.from_pretrained("squeezebert/squeezebert-mnli" ) __snake_case = torch.tensor([[1, 29_414, 232, 328, 740, 1_140, 12_695, 69, 13, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 3) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor([[0.6401, -0.0349, -0.6041]] ) self.assertTrue(torch.allclose(a_ , a_ , atol=1e-4 ) )
69
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
1
'''simple docstring''' import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets a : List[Any] = '''\ @inproceedings{kakwani2020indicnlpsuite, title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}}, author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar}, year={2020}, booktitle={Findings of EMNLP}, } ''' a : Union[str, Any] = '''\ IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te. ''' a : str = ''' Compute IndicGLUE evaluation metric associated to each IndicGLUE dataset. Args: predictions: list of predictions to score (as int64), except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32). references: list of ground truth labels corresponding to the predictions (as int64), except for \'cvit-mkb-clsr\' where each reference is a vector (of float32). Returns: depending on the IndicGLUE subset, one or several of: "accuracy": Accuracy "f1": F1 score "precision": Precision@10 Examples: >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\') >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'precision@10\': 1.0} ''' def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : str ) -> Tuple: return float((preds == labels).mean() ) def __UpperCAmelCase ( _UpperCAmelCase : Dict , _UpperCAmelCase : List[str] ) -> int: __snake_case = simple_accuracy(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = float(fa_score(y_true=_UpperCAmelCase , y_pred=_UpperCAmelCase ) ) return { "accuracy": acc, "f1": fa, } def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : str ) -> Union[str, Any]: __snake_case = np.array(_UpperCAmelCase ) __snake_case = np.array(_UpperCAmelCase ) __snake_case = en_sentvecs.shape[0] # mean centering __snake_case = en_sentvecs - np.mean(_UpperCAmelCase , axis=0 ) __snake_case = in_sentvecs - np.mean(_UpperCAmelCase , axis=0 ) __snake_case = cdist(_UpperCAmelCase , _UpperCAmelCase , "cosine" ) __snake_case = np.array(range(_UpperCAmelCase ) ) __snake_case = sim.argsort(axis=1 )[:, :10] __snake_case = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def A ( self : Optional[Any] ): """simple docstring""" if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]" ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" ) if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32" ) ), "references": datasets.Value("int64" ) if self.config_name != "cvit-mkb-clsr" else datasets.Sequence(datasets.Value("float32" ) ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" if self.config_name != "cvit-mkb-clsr" else None , ) def A ( self : str , a_ : Any , a_ : Optional[int] ): """simple docstring""" if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(a_ , a_ )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(a_ , a_ ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(a_ , a_ )} else: raise KeyError( "You should supply a configuration name selected in " "[\"wnli\", \"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", " "\"cvit-mkb-clsr\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\", " "\"wiki-ner\"]" )
69
'''simple docstring''' import json import os import torch from diffusers import UNetaDModel os.makedirs('''hub/hopper-medium-v2/unet/hor32''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/unet/hor128''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/value_function''', exist_ok=True) def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> str: if hor == 1_28: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D") elif hor == 32: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 64, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D", "UpResnetBlock1D") __snake_case = torch.load(F'''/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch''' ) __snake_case = model.state_dict() __snake_case = { "down_block_types": down_block_types, "block_out_channels": block_out_channels, "up_block_types": up_block_types, "layers_per_block": 1, "use_timestep_embedding": True, "out_block_type": "OutConv1DBlock", "norm_num_groups": 8, "downsample_each_block": False, "in_channels": 14, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "flip_sin_to_cos": False, "freq_shift": 1, "sample_size": 6_55_36, "mid_block_type": "MidResTemporalBlock1D", "act_fn": "mish", } __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , F'''hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin''' ) with open(F'''hub/hopper-medium-v2/unet/hor{hor}/config.json''' , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) def __UpperCAmelCase ( ) -> List[Any]: __snake_case = { "in_channels": 14, "down_block_types": ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D"), "up_block_types": (), "out_block_type": "ValueFunction", "mid_block_type": "ValueFunctionMidBlock1D", "block_out_channels": (32, 64, 1_28, 2_56), "layers_per_block": 1, "downsample_each_block": True, "sample_size": 6_55_36, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "use_timestep_embedding": True, "flip_sin_to_cos": False, "freq_shift": 1, "norm_num_groups": 8, "act_fn": "mish", } __snake_case = torch.load("/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch" ) __snake_case = model __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , "hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin" ) with open("hub/hopper-medium-v2/value_function/config.json" , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": unet(32) # unet(128) value_function()
69
1
'''simple docstring''' import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml a : str = NewType('''DataClass''', Any) a : Union[str, Any] = NewType('''DataClassType''', Any) def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Any: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F'''Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).''' ) def __UpperCAmelCase ( _UpperCAmelCase : list ) -> Callable[[str], Any]: __snake_case = {str(_UpperCAmelCase ): choice for choice in choices} return lambda _UpperCAmelCase : str_to_choice.get(_UpperCAmelCase , _UpperCAmelCase ) def __UpperCAmelCase ( *, _UpperCAmelCase : Union[str, List[str]] = None , _UpperCAmelCase : str = None , _UpperCAmelCase : Any = dataclasses.MISSING , _UpperCAmelCase : Callable[[], Any] = dataclasses.MISSING , _UpperCAmelCase : dict = None , **_UpperCAmelCase : Optional[int] , ) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls __snake_case = {} if aliases is not None: __snake_case = aliases if help is not None: __snake_case = help return dataclasses.field(metadata=_UpperCAmelCase , default=_UpperCAmelCase , default_factory=_UpperCAmelCase , **_UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = 42 def __init__( self : Dict , a_ : Union[DataClassType, Iterable[DataClassType]] , **a_ : List[Any] ): """simple docstring""" if "formatter_class" not in kwargs: __snake_case = ArgumentDefaultsHelpFormatter super().__init__(**a_ ) if dataclasses.is_dataclass(a_ ): __snake_case = [dataclass_types] __snake_case = list(a_ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(a_ ) @staticmethod def A ( a_ : ArgumentParser , a_ : dataclasses.Field ): """simple docstring""" __snake_case = f'''--{field.name}''' __snake_case = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , a_ ): raise RuntimeError( "Unresolved type detected, which should have been done with the help of " "`typing.get_type_hints` method by default" ) __snake_case = kwargs.pop("aliases" , [] ) if isinstance(a_ , a_ ): __snake_case = [aliases] __snake_case = getattr(field.type , "__origin__" , field.type ) if origin_type is Union or (hasattr(a_ , "UnionType" ) and isinstance(a_ , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(a_ ) not in field.type.__args__ ): raise ValueError( "Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because" " the argument parser only supports one type per argument." f''' Problem encountered in field \'{field.name}\'.''' ) if type(a_ ) not in field.type.__args__: # filter `str` in Union __snake_case = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] __snake_case = getattr(field.type , "__origin__" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) __snake_case = ( field.type.__args__[0] if isinstance(a_ , field.type.__args__[1] ) else field.type.__args__[1] ) __snake_case = getattr(field.type , "__origin__" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) __snake_case = {} if origin_type is Literal or (isinstance(field.type , a_ ) and issubclass(field.type , a_ )): if origin_type is Literal: __snake_case = field.type.__args__ else: __snake_case = [x.value for x in field.type] __snake_case = make_choice_type_function(kwargs["choices"] ) if field.default is not dataclasses.MISSING: __snake_case = field.default else: __snake_case = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument __snake_case = copy(a_ ) # Hack because type=bool in argparse does not behave as we want. __snake_case = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. __snake_case = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way __snake_case = default # This tells argparse we accept 0 or 1 value after --field_name __snake_case = "?" # This is the value that will get picked if we do --field_name (without value) __snake_case = True elif isclass(a_ ) and issubclass(a_ , a_ ): __snake_case = field.type.__args__[0] __snake_case = "+" if field.default_factory is not dataclasses.MISSING: __snake_case = field.default_factory() elif field.default is dataclasses.MISSING: __snake_case = True else: __snake_case = field.type if field.default is not dataclasses.MISSING: __snake_case = field.default elif field.default_factory is not dataclasses.MISSING: __snake_case = field.default_factory() else: __snake_case = True parser.add_argument(a_ , *a_ , **a_ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): __snake_case = False parser.add_argument(f'''--no_{field.name}''' , action="store_false" , dest=field.name , **a_ ) def A ( self : List[str] , a_ : DataClassType ): """simple docstring""" if hasattr(a_ , "_argument_group_name" ): __snake_case = self.add_argument_group(dtype._argument_group_name ) else: __snake_case = self try: __snake_case = get_type_hints(a_ ) except NameError: raise RuntimeError( f'''Type resolution failed for {dtype}. Try declaring the class in global scope or ''' "removing line of `from __future__ import annotations` which opts in Postponed " "Evaluation of Annotations (PEP 563)" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(a_ ): __snake_case = ".".join(map(a_ , sys.version_info[:3] ) ) raise RuntimeError( f'''Type resolution failed for {dtype} on Python {python_version}. Try removing ''' "line of `from __future__ import annotations` which opts in union types as " "`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To " "support Python versions that lower than 3.10, you need to use " "`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of " "`X | None`." ) from ex raise for field in dataclasses.fields(a_ ): if not field.init: continue __snake_case = type_hints[field.name] self._parse_dataclass_field(a_ , a_ ) def A ( self : Union[str, Any] , a_ : Optional[Any]=None , a_ : Tuple=False , a_ : Any=True , a_ : Optional[Any]=None , a_ : Dict=None , ): """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): __snake_case = [] if args_filename: args_files.append(Path(a_ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(".args" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values __snake_case = ArgumentParser() args_file_parser.add_argument(a_ , type=a_ , action="append" ) # Use only remaining args for further parsing (remove the args_file_flag) __snake_case , __snake_case = args_file_parser.parse_known_args(args=a_ ) __snake_case = vars(a_ ).get(args_file_flag.lstrip("-" ) , a_ ) if cmd_args_file_paths: args_files.extend([Path(a_ ) for p in cmd_args_file_paths] ) __snake_case = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last __snake_case = file_args + args if args is not None else file_args + sys.argv[1:] __snake_case , __snake_case = self.parse_known_args(args=a_ ) __snake_case = [] for dtype in self.dataclass_types: __snake_case = {f.name for f in dataclasses.fields(a_ ) if f.init} __snake_case = {k: v for k, v in vars(a_ ).items() if k in keys} for k in keys: delattr(a_ , a_ ) __snake_case = dtype(**a_ ) outputs.append(a_ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(a_ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f'''Some specified arguments are not used by the HfArgumentParser: {remaining_args}''' ) return (*outputs,) def A ( self : Optional[Any] , a_ : Dict[str, Any] , a_ : bool = False ): """simple docstring""" __snake_case = set(args.keys() ) __snake_case = [] for dtype in self.dataclass_types: __snake_case = {f.name for f in dataclasses.fields(a_ ) if f.init} __snake_case = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) __snake_case = dtype(**a_ ) outputs.append(a_ ) if not allow_extra_keys and unused_keys: raise ValueError(f'''Some keys are not used by the HfArgumentParser: {sorted(a_ )}''' ) return tuple(a_ ) def A ( self : Union[str, Any] , a_ : str , a_ : bool = False ): """simple docstring""" with open(Path(a_ ) , encoding="utf-8" ) as open_json_file: __snake_case = json.loads(open_json_file.read() ) __snake_case = self.parse_dict(a_ , allow_extra_keys=a_ ) return tuple(a_ ) def A ( self : int , a_ : str , a_ : bool = False ): """simple docstring""" __snake_case = self.parse_dict(yaml.safe_load(Path(a_ ).read_text() ) , allow_extra_keys=a_ ) return tuple(a_ )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 1_00_00_00 ) -> int: __snake_case = 1 __snake_case = 1 __snake_case = {1: 1} for inputa in range(2 , _UpperCAmelCase ): __snake_case = 0 __snake_case = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __snake_case = (3 * number) + 1 counter += 1 if inputa not in counters: __snake_case = counter if counter > pre_counter: __snake_case = inputa __snake_case = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
69
1
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE__ : def __init__( self : Optional[Any] , a_ : Optional[int] , a_ : Optional[int]=12 , a_ : Dict=7 , a_ : str=True , a_ : List[Any]=True , a_ : Any=True , a_ : List[str]=99 , a_ : int=32 , a_ : Any=32 , a_ : Union[str, Any]=2 , a_ : Optional[int]=4 , a_ : Any=37 , a_ : List[str]=0.1 , a_ : int=0.1 , a_ : str=512 , a_ : Optional[Any]=0.02 , a_ : List[Any]=0 , a_ : Any=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = projection_dim __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = dropout __snake_case = attention_dropout __snake_case = max_position_embeddings __snake_case = initializer_range __snake_case = scope __snake_case = bos_token_id def A ( self : Tuple ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: __snake_case = input_mask.numpy() __snake_case , __snake_case = input_mask.shape __snake_case = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a_ ): __snake_case = 1 __snake_case = 0 __snake_case = self.get_config() return config, input_ids, tf.convert_to_tensor(a_ ) def A ( self : List[str] ): """simple docstring""" return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def A ( self : Optional[int] , a_ : Union[str, Any] , a_ : List[str] , a_ : str ): """simple docstring""" __snake_case = TFBlipTextModel(config=a_ ) __snake_case = model(a_ , attention_mask=a_ , training=a_ ) __snake_case = model(a_ , training=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : str ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = (TFBlipTextModel,) if is_tf_available() else () __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = BlipTextModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Dict ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : Tuple ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : str ): """simple docstring""" pass def A ( self : List[Any] ): """simple docstring""" pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def A ( self : str ): """simple docstring""" pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def A ( self : Optional[Any] ): """simple docstring""" pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def A ( self : int ): """simple docstring""" pass @slow def A ( self : Optional[int] ): """simple docstring""" for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = TFBlipTextModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def A ( self : Optional[int] , a_ : int=True ): """simple docstring""" super().test_pt_tf_model_equivalence(allow_missing_keys=a_ )
69
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """SpeechT5FeatureExtractor""" __SCREAMING_SNAKE_CASE = """SpeechT5Tokenizer""" def __init__( self : List[Any] , a_ : str , a_ : str ): """simple docstring""" super().__init__(a_ , a_ ) def __call__( self : Dict , *a_ : Tuple , **a_ : List[str] ): """simple docstring""" __snake_case = kwargs.pop("audio" , a_ ) __snake_case = kwargs.pop("text" , a_ ) __snake_case = kwargs.pop("text_target" , a_ ) __snake_case = kwargs.pop("audio_target" , a_ ) __snake_case = kwargs.pop("sampling_rate" , a_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __snake_case = self.feature_extractor(a_ , *a_ , sampling_rate=a_ , **a_ ) elif text is not None: __snake_case = self.tokenizer(a_ , **a_ ) else: __snake_case = None if audio_target is not None: __snake_case = self.feature_extractor(audio_target=a_ , *a_ , sampling_rate=a_ , **a_ ) __snake_case = targets["input_values"] elif text_target is not None: __snake_case = self.tokenizer(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : str , **a_ : Dict ): """simple docstring""" __snake_case = kwargs.pop("input_values" , a_ ) __snake_case = kwargs.pop("input_ids" , a_ ) __snake_case = kwargs.pop("labels" , a_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) elif input_ids is not None: __snake_case = self.tokenizer.pad(a_ , **a_ ) else: __snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(a_ , a_ ) and "input_ids" in labels[0]): __snake_case = self.tokenizer.pad(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = self.feature_extractor.feature_size __snake_case = self.feature_extractor.num_mel_bins __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) __snake_case = feature_size_hack __snake_case = targets["input_values"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : Any , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : Optional[int] , *a_ : Union[str, Any] , **a_ : str ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ )
69
1
'''simple docstring''' from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
69
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : Optional[Any] , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __snake_case = input_file.read() __snake_case = regexp.search(a_ ) return match def A ( self : Any , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __snake_case = regexp.finditer(a_ ) __snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A ( self : Optional[int] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a_ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(a_ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 __snake_case , __snake_case = 1, 1 for _ in range(number_of_steps - 1 ): __snake_case , __snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : Dict = {'''vocab_file''': '''sentencepiece.model'''} a : Tuple = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } a : str = { '''google/rembert''': 256, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[Any] , a_ : int , a_ : Any=False , a_ : List[Any]=True , a_ : List[Any]=True , a_ : List[Any]="[CLS]" , a_ : List[Any]="[SEP]" , a_ : List[Any]="[UNK]" , a_ : str="[SEP]" , a_ : List[str]="[PAD]" , a_ : Optional[int]="[CLS]" , a_ : List[str]="[MASK]" , **a_ : str , ): """simple docstring""" super().__init__( do_lower_case=a_ , remove_space=a_ , keep_accents=a_ , bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , pad_token=a_ , cls_token=a_ , mask_token=a_ , **a_ , ) __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(a_ ) @property def A ( self : Optional[Any] ): """simple docstring""" return len(self.sp_model ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ): """simple docstring""" __snake_case = self.__dict__.copy() __snake_case = None return state def __setstate__( self : str , a_ : Optional[int] ): """simple docstring""" __snake_case = d __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def A ( self : Tuple , a_ : Optional[int] , a_ : int=False ): """simple docstring""" __snake_case = self.sp_model.EncodeAsPieces(a_ ) return pieces def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" return self.sp_model.PieceToId(a_ ) def A ( self : Optional[Any] , a_ : List[str] ): """simple docstring""" return self.sp_model.IdToPiece(a_ ) def A ( self : Optional[Any] , a_ : int ): """simple docstring""" __snake_case = self.sp_model.decode_pieces(a_ ) return out_string def A ( self : Union[str, Any] , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a_ )) + [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1] def A ( self : Tuple , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A ( self : List[Any] , a_ : str , a_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a_ ): logger.error("Vocabulary path ({}) should be a directory".format(a_ ) ) return __snake_case = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ): copyfile(self.vocab_file , a_ ) return (out_vocab_file,)
69
1
'''simple docstring''' import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class SCREAMING_SNAKE_CASE__ : def __init__( self : Optional[int] , a_ : str = "cpu" , a_ : str = "openai/clip-vit-large-patch14" ): """simple docstring""" __snake_case = device __snake_case = CLIPTokenizerFast.from_pretrained(a_ ) __snake_case = [0.48145466, 0.4578275, 0.40821073] __snake_case = [0.26862954, 0.26130258, 0.27577711] __snake_case = torchvision.transforms.Normalize(self.image_mean , self.image_std ) __snake_case = torchvision.transforms.Resize(224 ) __snake_case = torchvision.transforms.CenterCrop(224 ) def A ( self : Tuple , a_ : int ): """simple docstring""" __snake_case = self.resize(a_ ) __snake_case = self.center_crop(a_ ) __snake_case = self.normalize(a_ ) return images def __call__( self : List[str] , a_ : Any=None , a_ : Any=None , **a_ : Optional[int] ): """simple docstring""" __snake_case = self.tokenizer(text=a_ , **a_ ) __snake_case = self.preprocess_img(a_ ) __snake_case = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : List[str] , a_ : int=10 , a_ : str=0.01 , a_ : str=None , a_ : Optional[int]=None , a_ : Optional[Any]=None , a_ : Tuple=None , a_ : Tuple=None , a_ : List[str]=None , a_ : List[str]=False , a_ : List[str]=True , a_ : List[Any]="image" , a_ : str=True , a_ : List[str]=False , a_ : Any=False , a_ : Dict=False , ): """simple docstring""" super().__init__() __snake_case = None __snake_case = device if device else get_device() if vqgan: __snake_case = vqgan else: __snake_case = load_vqgan(self.device , conf_path=a_ , ckpt_path=a_ ) self.vqgan.eval() if clip: __snake_case = clip else: __snake_case = CLIPModel.from_pretrained("openai/clip-vit-base-patch32" ) self.clip.to(self.device ) __snake_case = ProcessorGradientFlow(device=self.device ) __snake_case = iterations __snake_case = lr __snake_case = log __snake_case = make_grid __snake_case = return_val __snake_case = quantize __snake_case = self.vqgan.decoder.z_shape def A ( self : Union[str, Any] , a_ : Dict=None , a_ : Dict=None , a_ : Any=5 , a_ : int=True ): """simple docstring""" __snake_case = [] if output_path is None: __snake_case = "./animation.gif" if input_path is None: __snake_case = self.save_path __snake_case = sorted(glob(input_path + "/*" ) ) if not len(a_ ): raise ValueError( "No images found in save path, aborting (did you pass save_intermediate=True to the generate" " function?)" ) if len(a_ ) == 1: print("Only one image found in save path, (did you pass save_intermediate=True to the generate function?)" ) __snake_case = total_duration / len(a_ ) __snake_case = [frame_duration] * len(a_ ) if extend_frames: __snake_case = 1.5 __snake_case = 3 for file_name in paths: if file_name.endswith(".png" ): images.append(imageio.imread(a_ ) ) imageio.mimsave(a_ , a_ , duration=a_ ) print(f'''gif saved to {output_path}''' ) def A ( self : Any , a_ : Union[str, Any]=None , a_ : Union[str, Any]=None ): """simple docstring""" if not (path or img): raise ValueError("Input either path or tensor" ) if img is not None: raise NotImplementedError __snake_case = preprocess(Image.open(a_ ) , target_image_size=256 ).to(self.device ) __snake_case = preprocess_vqgan(a_ ) __snake_case , *__snake_case = self.vqgan.encode(a_ ) return z def A ( self : List[str] , a_ : Dict ): """simple docstring""" __snake_case = self.latent.detach().requires_grad_() __snake_case = base_latent + transform_vector if self.quantize: __snake_case , *__snake_case = self.vqgan.quantize(a_ ) else: __snake_case = trans_latent return self.vqgan.decode(a_ ) def A ( self : str , a_ : Any , a_ : int , a_ : str=None ): """simple docstring""" __snake_case = self.clip_preprocessor(text=a_ , images=a_ , return_tensors="pt" , padding=a_ ) __snake_case = self.clip(**a_ ) __snake_case = clip_outputs.logits_per_image if weights is not None: __snake_case = similarity_logits * weights return similarity_logits.sum() def A ( self : int , a_ : List[str] , a_ : Union[str, Any] , a_ : Dict ): """simple docstring""" __snake_case = self._get_clip_similarity(pos_prompts["prompts"] , a_ , weights=(1 / pos_prompts["weights"]) ) if neg_prompts: __snake_case = self._get_clip_similarity(neg_prompts["prompts"] , a_ , weights=neg_prompts["weights"] ) else: __snake_case = torch.tensor([1] , device=self.device ) __snake_case = -torch.log(a_ ) + torch.log(a_ ) return loss def A ( self : str , a_ : List[str] , a_ : List[Any] , a_ : Union[str, Any] ): """simple docstring""" __snake_case = torch.randn_like(self.latent , requires_grad=a_ , device=self.device ) __snake_case = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() __snake_case = self._add_vector(a_ ) __snake_case = loop_post_process(a_ ) __snake_case = self._get_CLIP_loss(a_ , a_ , a_ ) print("CLIP loss" , a_ ) if self.log: wandb.log({"CLIP Loss": clip_loss} ) clip_loss.backward(retain_graph=a_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def A ( self : Union[str, Any] , a_ : Dict , a_ : int , a_ : Tuple ): """simple docstring""" wandb.init(reinit=a_ , project="face-editor" ) wandb.config.update({"Positive Prompts": positive_prompts} ) wandb.config.update({"Negative Prompts": negative_prompts} ) wandb.config.update({"lr": self.lr, "iterations": self.iterations} ) if image_path: __snake_case = Image.open(a_ ) __snake_case = image.resize((256, 256) ) wandb.log("Original Image" , wandb.Image(a_ ) ) def A ( self : Tuple , a_ : Dict ): """simple docstring""" if not prompts: return [] __snake_case = [] __snake_case = [] if isinstance(a_ , a_ ): __snake_case = [prompt.strip() for prompt in prompts.split("|" )] for prompt in prompts: if isinstance(a_ , (tuple, list) ): __snake_case = prompt[0] __snake_case = float(prompt[1] ) elif ":" in prompt: __snake_case , __snake_case = prompt.split(":" ) __snake_case = float(a_ ) else: __snake_case = prompt __snake_case = 1.0 processed_prompts.append(a_ ) weights.append(a_ ) return { "prompts": processed_prompts, "weights": torch.tensor(a_ , device=self.device ), } def A ( self : List[Any] , a_ : Optional[Any] , a_ : Dict=None , a_ : Optional[Any]=None , a_ : List[Any]=True , a_ : List[str]=False , a_ : Any=True , a_ : Union[str, Any]=True , a_ : str=None , ): """simple docstring""" if image_path: __snake_case = self._get_latent(a_ ) else: __snake_case = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(a_ , a_ , a_ ) assert pos_prompts, "You must provide at least one positive prompt." __snake_case = self.process_prompts(a_ ) __snake_case = self.process_prompts(a_ ) if save_final and save_path is None: __snake_case = os.path.join("./outputs/" , "_".join(pos_prompts["prompts"] ) ) if not os.path.exists(a_ ): os.makedirs(a_ ) else: __snake_case = save_path + "_" + get_timestamp() os.makedirs(a_ ) __snake_case = save_path __snake_case = self.vqgan.decode(self.latent )[0] if show_intermediate: print("Original Image" ) show_pil(custom_to_pil(a_ ) ) __snake_case = loop_post_process(a_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(a_ , a_ , a_ ) ): if show_intermediate: show_pil(a_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}.png''' ) ) if self.log: wandb.log({"Image": wandb.Image(a_ )} ) if show_final: show_pil(a_ ) if save_final: transformed_img.save(os.path.join(self.save_path , f'''iter_{iter:03d}_final.png''' ) )
69
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ): """simple docstring""" try: __snake_case = tempfile.mktemp() with open(a_ , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , a_ ) __snake_case = AlbertTokenizer.from_pretrained(a_ ) finally: os.remove(a_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , a_ ) __snake_case = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A ( self : str ): """simple docstring""" __snake_case = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : List[Any] ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a_ , repo_id="test-tokenizer" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a_ , repo_id="valid_org/test-tokenizer-org" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A ( self : List[str] ): """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizerFast.from_pretrained(a_ ) bert_tokenizer.save_pretrained(a_ ) __snake_case = CustomTokenizerFast.from_pretrained(a_ ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a_ , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A ( self : str ): """simple docstring""" __snake_case = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def A ( self : List[Any] ): """simple docstring""" __snake_case = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : str ): """simple docstring""" __snake_case = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def A ( self : Tuple ): """simple docstring""" __snake_case = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def A ( self : Any ): """simple docstring""" __snake_case = Trie() __snake_case = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a_ , ["AB", "C"] )
69
1
'''simple docstring''' import numpy as np def __UpperCAmelCase ( _UpperCAmelCase : np.array ) -> np.array: return 1 / (1 + np.exp(-vector )) def __UpperCAmelCase ( _UpperCAmelCase : np.array ) -> np.array: return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 __snake_case , __snake_case = 1, 1 for _ in range(number_of_steps - 1 ): __snake_case , __snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' from .configuration_bert_masked import MaskedBertConfig from .modeling_bert_masked import ( MaskedBertForMultipleChoice, MaskedBertForQuestionAnswering, MaskedBertForSequenceClassification, MaskedBertForTokenClassification, MaskedBertModel, ) from .modules import *
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str ) -> str: return " ".join( "".join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
69
1
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType a : Optional[int] = logging.get_logger(__name__) a : Optional[Any] = { '''microsoft/deberta-v2-xlarge''': '''https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xxlarge''': '''https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json''' ), '''microsoft/deberta-v2-xxlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """deberta-v2""" def __init__( self : str , a_ : List[str]=128_100 , a_ : Any=1_536 , a_ : Tuple=24 , a_ : Optional[Any]=24 , a_ : Optional[Any]=6_144 , a_ : Tuple="gelu" , a_ : Optional[int]=0.1 , a_ : Tuple=0.1 , a_ : Optional[int]=512 , a_ : int=0 , a_ : List[Any]=0.02 , a_ : List[str]=1e-7 , a_ : str=False , a_ : Optional[Any]=-1 , a_ : Union[str, Any]=0 , a_ : Dict=True , a_ : Union[str, Any]=None , a_ : Any=0 , a_ : Optional[Any]="gelu" , **a_ : Optional[Any] , ): """simple docstring""" super().__init__(**a_ ) __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = initializer_range __snake_case = relative_attention __snake_case = max_relative_positions __snake_case = pad_token_id __snake_case = position_biased_input # Backwards compatibility if type(a_ ) == str: __snake_case = [x.strip() for x in pos_att_type.lower().split("|" )] __snake_case = pos_att_type __snake_case = vocab_size __snake_case = layer_norm_eps __snake_case = kwargs.get("pooler_hidden_size" , a_ ) __snake_case = pooler_dropout __snake_case = pooler_hidden_act class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @property def A ( self : int ): """simple docstring""" if self.task == "multiple-choice": __snake_case = {0: "batch", 1: "choice", 2: "sequence"} else: __snake_case = {0: "batch", 1: "sequence"} if self._config.type_vocab_size > 0: return OrderedDict( [("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis)] ) else: return OrderedDict([("input_ids", dynamic_axis), ("attention_mask", dynamic_axis)] ) @property def A ( self : List[str] ): """simple docstring""" return 12 def A ( self : int , a_ : Union["PreTrainedTokenizerBase", "FeatureExtractionMixin"] , a_ : int = -1 , a_ : int = -1 , a_ : int = -1 , a_ : bool = False , a_ : Optional["TensorType"] = None , a_ : int = 3 , a_ : int = 40 , a_ : int = 40 , a_ : "PreTrainedTokenizerBase" = None , ): """simple docstring""" __snake_case = super().generate_dummy_inputs(preprocessor=a_ , framework=a_ ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
69
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : Any , a_ : Union[str, Any]=13 , a_ : Any=7 , a_ : Any=True , a_ : Dict=True , a_ : Union[str, Any]=False , a_ : Tuple=True , a_ : str=99 , a_ : Tuple=64 , a_ : Tuple=5 , a_ : Union[str, Any]=4 , a_ : Dict=64 , a_ : Union[str, Any]="gelu" , a_ : Dict=0.1 , a_ : List[str]=0.1 , a_ : Dict=512 , a_ : Tuple=16 , a_ : str=2 , a_ : Any=0.02 , a_ : List[Any]=3 , a_ : Tuple=4 , a_ : Optional[int]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : int ): """simple docstring""" return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def A ( self : str ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def A ( self : Tuple , a_ : int , a_ : str , a_ : Optional[int] , a_ : List[Any] , a_ : str , a_ : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Any , a_ : int , a_ : Tuple , a_ : str , a_ : int , a_ : str , a_ : List[Any] ): """simple docstring""" __snake_case = MPNetForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Any , a_ : Any , a_ : int , a_ : Union[str, Any] , a_ : Dict , a_ : Optional[Any] , a_ : Any ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Optional[Any] , a_ : Any , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = MPNetForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict , a_ : List[str] , a_ : str , a_ : Union[str, Any] , a_ : str , a_ : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def A ( self : List[Any] ): """simple docstring""" __snake_case = MPNetModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel.from_pretrained("microsoft/mpnet-base" ) __snake_case = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a_ , atol=1e-4 ) )
69
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices a : Optional[int] = logging.get_logger(__name__) a : Tuple = { '''microsoft/focalnet-tiny''': '''https://huggingface.co/microsoft/focalnet-tiny/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """focalnet""" def __init__( self : int , a_ : Optional[int]=224 , a_ : int=4 , a_ : Dict=3 , a_ : Any=96 , a_ : str=False , a_ : Optional[Any]=[192, 384, 768, 768] , a_ : Tuple=[2, 2, 6, 2] , a_ : Dict=[2, 2, 2, 2] , a_ : List[Any]=[3, 3, 3, 3] , a_ : Union[str, Any]="gelu" , a_ : Union[str, Any]=4.0 , a_ : List[str]=0.0 , a_ : int=0.1 , a_ : List[Any]=False , a_ : Tuple=1e-4 , a_ : Tuple=False , a_ : Dict=False , a_ : List[Any]=False , a_ : List[str]=0.02 , a_ : Dict=1e-5 , a_ : Union[str, Any]=32 , a_ : List[str]=None , a_ : str=None , **a_ : Union[str, Any] , ): """simple docstring""" super().__init__(**a_ ) __snake_case = image_size __snake_case = patch_size __snake_case = num_channels __snake_case = embed_dim __snake_case = use_conv_embed __snake_case = hidden_sizes __snake_case = depths __snake_case = focal_levels __snake_case = focal_windows __snake_case = hidden_act __snake_case = mlp_ratio __snake_case = hidden_dropout_prob __snake_case = drop_path_rate __snake_case = use_layerscale __snake_case = layerscale_value __snake_case = use_post_layernorm __snake_case = use_post_layernorm_in_modulation __snake_case = normalize_modulator __snake_case = initializer_range __snake_case = layer_norm_eps __snake_case = encoder_stride __snake_case = ["stem"] + [f'''stage{idx}''' for idx in range(1 , len(self.depths ) + 1 )] __snake_case , __snake_case = get_aligned_output_features_output_indices( out_features=a_ , out_indices=a_ , stage_names=self.stage_names )
69
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Optional[int]: return 1 / (1 + np.exp(-z )) def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict ) -> List[str]: return (-y * np.log(_UpperCAmelCase ) - (1 - y) * np.log(1 - h )).mean() def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any] ) -> Optional[Any]: __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) return np.sum(y * scores - np.log(1 + np.exp(_UpperCAmelCase ) ) ) def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str]=7_00_00 ) -> Union[str, Any]: __snake_case = np.zeros(x.shape[1] ) for iterations in range(_UpperCAmelCase ): __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = np.dot(x.T , h - y ) / y.size __snake_case = theta - alpha * gradient # updating the weights __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = cost_function(_UpperCAmelCase , _UpperCAmelCase ) if iterations % 1_00 == 0: print(F'''loss: {j} \t''' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": a : int = datasets.load_iris() a : int = iris.data[:, :2] a : Optional[Any] = (iris.target != 0) * 1 a : Tuple = 0.1 a : List[str] = logistic_reg(alpha, x, y, max_iterations=70_000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> Union[str, Any]: return sigmoid_function( np.dot(_UpperCAmelCase , _UpperCAmelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((a) , (a)) : Any = (x[:, 0].min(), x[:, 0].max()) ((a) , (a)) : Any = (x[:, 1].min(), x[:, 1].max()) ((a) , (a)) : Any = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) a : Optional[Any] = np.c_[xxa.ravel(), xxa.ravel()] a : List[Any] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
69
1
'''simple docstring''' from __future__ import annotations from random import random from typing import Generic, TypeVar a : Any = TypeVar('''KT''') a : Optional[int] = TypeVar('''VT''') class SCREAMING_SNAKE_CASE__ ( Generic[KT, VT] ): def __init__( self : int , a_ : KT | str = "root" , a_ : VT | None = None ): """simple docstring""" __snake_case = key __snake_case = value __snake_case = [] def __repr__( self : Union[str, Any] ): """simple docstring""" return f'''Node({self.key}: {self.value})''' @property def A ( self : Tuple ): """simple docstring""" return len(self.forward ) class SCREAMING_SNAKE_CASE__ ( Generic[KT, VT] ): def __init__( self : Dict , a_ : float = 0.5 , a_ : int = 16 ): """simple docstring""" __snake_case = Node[KT, VT]() __snake_case = 0 __snake_case = p __snake_case = max_level def __str__( self : int ): """simple docstring""" __snake_case = list(self ) if len(a_ ) == 0: return f'''SkipList(level={self.level})''' __snake_case = max((len(str(a_ ) ) for item in items) , default=4 ) __snake_case = max(a_ , 4 ) + 4 __snake_case = self.head __snake_case = [] __snake_case = node.forward.copy() lines.append(f'''[{node.key}]'''.ljust(a_ , "-" ) + "* " * len(a_ ) ) lines.append(" " * label_size + "| " * len(a_ ) ) while len(node.forward ) != 0: __snake_case = node.forward[0] lines.append( f'''[{node.key}]'''.ljust(a_ , "-" ) + " ".join(str(n.key ) if n.key == node.key else "|" for n in forwards ) ) lines.append(" " * label_size + "| " * len(a_ ) ) __snake_case = node.forward lines.append("None".ljust(a_ ) + "* " * len(a_ ) ) return f'''SkipList(level={self.level})\n''' + "\n".join(a_ ) def __iter__( self : int ): """simple docstring""" __snake_case = self.head while len(node.forward ) != 0: yield node.forward[0].key __snake_case = node.forward[0] def A ( self : List[Any] ): """simple docstring""" __snake_case = 1 while random() < self.p and level < self.max_level: level += 1 return level def A ( self : str , a_ : Tuple ): """simple docstring""" __snake_case = [] __snake_case = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: __snake_case = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(a_ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def A ( self : Union[str, Any] , a_ : KT ): """simple docstring""" __snake_case , __snake_case = self._locate_node(a_ ) if node is not None: for i, update_node in enumerate(a_ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: __snake_case = node.forward[i] else: __snake_case = update_node.forward[:i] def A ( self : Tuple , a_ : KT , a_ : VT ): """simple docstring""" __snake_case , __snake_case = self._locate_node(a_ ) if node is not None: __snake_case = value else: __snake_case = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 , a_ ): update_vector.append(self.head ) __snake_case = level __snake_case = Node(a_ , a_ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(a_ ) else: __snake_case = new_node def A ( self : Any , a_ : VT ): """simple docstring""" __snake_case , __snake_case = self._locate_node(a_ ) if node is not None: return node.value return None def __UpperCAmelCase ( ) -> Optional[int]: __snake_case = SkipList() skip_list.insert("Key1" , 3 ) skip_list.insert("Key2" , 12 ) skip_list.insert("Key3" , 41 ) skip_list.insert("Key4" , -19 ) __snake_case = skip_list.head __snake_case = {} while node.level != 0: __snake_case = node.forward[0] __snake_case = node.value assert len(_UpperCAmelCase ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def __UpperCAmelCase ( ) -> int: __snake_case = SkipList() skip_list.insert("Key1" , 10 ) skip_list.insert("Key1" , 12 ) skip_list.insert("Key5" , 7 ) skip_list.insert("Key7" , 10 ) skip_list.insert("Key10" , 5 ) skip_list.insert("Key7" , 7 ) skip_list.insert("Key5" , 5 ) skip_list.insert("Key10" , 10 ) __snake_case = skip_list.head __snake_case = {} while node.level != 0: __snake_case = node.forward[0] __snake_case = node.value if len(_UpperCAmelCase ) != 4: print() assert len(_UpperCAmelCase ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def __UpperCAmelCase ( ) -> Optional[int]: __snake_case = SkipList() assert skip_list.find("Some key" ) is None def __UpperCAmelCase ( ) -> Union[str, Any]: __snake_case = SkipList() skip_list.insert("Key2" , 20 ) assert skip_list.find("Key2" ) == 20 skip_list.insert("Some Key" , 10 ) skip_list.insert("Key2" , 8 ) skip_list.insert("V" , 13 ) assert skip_list.find("Y" ) is None assert skip_list.find("Key2" ) == 8 assert skip_list.find("Some Key" ) == 10 assert skip_list.find("V" ) == 13 def __UpperCAmelCase ( ) -> List[str]: __snake_case = SkipList() skip_list.delete("Some key" ) assert len(skip_list.head.forward ) == 0 def __UpperCAmelCase ( ) -> Tuple: __snake_case = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("Key2" ) is None def __UpperCAmelCase ( ) -> str: __snake_case = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 14 ) skip_list.insert("Key2" , 15 ) skip_list.delete("V" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) == 14 assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("X" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) == 12 assert skip_list.find("Key2" ) == 15 skip_list.delete("Key1" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) == 15 skip_list.delete("Key2" ) assert skip_list.find("V" ) is None assert skip_list.find("X" ) is None assert skip_list.find("Key1" ) is None assert skip_list.find("Key2" ) is None def __UpperCAmelCase ( ) -> Optional[Any]: __snake_case = SkipList() skip_list.insert("Key1" , 12 ) skip_list.insert("V" , 13 ) skip_list.insert("X" , 1_42 ) skip_list.insert("Key2" , 15 ) skip_list.delete("X" ) def traverse_keys(_UpperCAmelCase : Any ): yield node.key for forward_node in node.forward: yield from traverse_keys(_UpperCAmelCase ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def __UpperCAmelCase ( ) -> Optional[Any]: def is_sorted(_UpperCAmelCase : Any ): return all(next_item >= item for item, next_item in zip(_UpperCAmelCase , lst[1:] ) ) __snake_case = SkipList() for i in range(10 ): skip_list.insert(_UpperCAmelCase , _UpperCAmelCase ) assert is_sorted(list(_UpperCAmelCase ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(_UpperCAmelCase ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(_UpperCAmelCase ) ) def __UpperCAmelCase ( ) -> int: for _ in range(1_00 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def __UpperCAmelCase ( ) -> Tuple: __snake_case = SkipList() skip_list.insert(2 , "2" ) skip_list.insert(4 , "4" ) skip_list.insert(6 , "4" ) skip_list.insert(4 , "5" ) skip_list.insert(8 , "4" ) skip_list.insert(9 , "4" ) skip_list.delete(4 ) print(_UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging a : int = logging.get_logger(__name__) a : Any = r''' Args: input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax or scores for each vocabulary token after SoftMax. kwargs (`Dict[str, Any]`, *optional*): Additional stopping criteria specific kwargs. Return: `bool`. `False` indicates we should continue, `True` indicates we should stop. ''' class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @add_start_docstrings(a_ ) def __call__( self : int , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Tuple ): """simple docstring""" raise NotImplementedError("StoppingCriteria needs to be subclassed" ) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Dict , a_ : int , a_ : Optional[int] = None ): """simple docstring""" __snake_case = max_length __snake_case = max_position_embeddings @add_start_docstrings(a_ ) def __call__( self : str , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Union[str, Any] ): """simple docstring""" __snake_case = input_ids.shape[-1] __snake_case = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( "This is a friendly reminder - the current text generation call will exceed the model's predefined " f'''maximum length ({self.max_position_embeddings}). Depending on the model, you may observe ''' "exceptions, performance degradation, or nothing at all." ) return is_done class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : List[str] , a_ : int , a_ : int ): """simple docstring""" warnings.warn( "The class `MaxNewTokensCriteria` is deprecated. " f'''Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` ''' "with `max_length = start_length + max_new_tokens` instead." , a_ , ) __snake_case = start_length __snake_case = max_new_tokens __snake_case = start_length + max_new_tokens @add_start_docstrings(a_ ) def __call__( self : Union[str, Any] , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Tuple ): """simple docstring""" return input_ids.shape[-1] >= self.max_length class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Tuple , a_ : float , a_ : Optional[float] = None ): """simple docstring""" __snake_case = max_time __snake_case = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(a_ ) def __call__( self : str , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : str ): """simple docstring""" return time.time() - self.initial_timestamp > self.max_time class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @add_start_docstrings(a_ ) def __call__( self : Tuple , a_ : torch.LongTensor , a_ : torch.FloatTensor , **a_ : Tuple ): """simple docstring""" return any(criteria(a_ , a_ ) for criteria in self ) @property def A ( self : Any ): """simple docstring""" for stopping_criterium in self: if isinstance(a_ , a_ ): return stopping_criterium.max_length elif isinstance(a_ , a_ ): return stopping_criterium.max_length return None def __UpperCAmelCase ( _UpperCAmelCase : StoppingCriteriaList , _UpperCAmelCase : int ) -> StoppingCriteriaList: __snake_case = stopping_criteria.max_length __snake_case = deepcopy(_UpperCAmelCase ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn("You set different `max_length` for stopping criteria and `max_length` parameter" , _UpperCAmelCase ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=_UpperCAmelCase ) ) return new_stopping_criteria
69
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
1
'''simple docstring''' import os from collections import deque import torch from torch.utils.data import Dataset class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : int , a_ : str="" , a_ : List[Any]="train" ): """simple docstring""" assert os.path.isdir(a_ ) __snake_case = [] __snake_case = os.listdir(a_ ) for story_filename in story_filenames_list: if "summary" in story_filename: continue __snake_case = os.path.join(a_ , a_ ) if not os.path.isfile(a_ ): continue self.documents.append(a_ ) def __len__( self : int ): """simple docstring""" return len(self.documents ) def __getitem__( self : List[Any] , a_ : Tuple ): """simple docstring""" __snake_case = self.documents[idx] __snake_case = document_path.split("/" )[-1] with open(a_ , encoding="utf-8" ) as source: __snake_case = source.read() __snake_case , __snake_case = process_story(a_ ) return document_name, story_lines, summary_lines def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> List[str]: __snake_case = list(filter(lambda _UpperCAmelCase : len(_UpperCAmelCase ) != 0 , [line.strip() for line in raw_story.split("\n" )] ) ) # for some unknown reason some lines miss a period, add it __snake_case = [_add_missing_period(_UpperCAmelCase ) for line in nonempty_lines] # gather article lines __snake_case = [] __snake_case = deque(_UpperCAmelCase ) while True: try: __snake_case = lines.popleft() if element.startswith("@highlight" ): break story_lines.append(_UpperCAmelCase ) except IndexError: # if "@highlight" is absent from the file we pop # all elements until there is None, raising an exception. return story_lines, [] # gather summary lines __snake_case = list(filter(lambda _UpperCAmelCase : not t.startswith("@highlight" ) , _UpperCAmelCase ) ) return story_lines, summary_lines def __UpperCAmelCase ( _UpperCAmelCase : Tuple ) -> Tuple: __snake_case = [".", "!", "?", "...", "'", "`", "\"", "\u2019", "\u2019", ")"] if line.startswith("@highlight" ): return line if line[-1] in END_TOKENS: return line return line + "." def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict , _UpperCAmelCase : Dict ) -> Dict: if len(_UpperCAmelCase ) > block_size: return sequence[:block_size] else: sequence.extend([pad_token_id] * (block_size - len(_UpperCAmelCase )) ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : Any ) -> Dict: __snake_case = torch.ones_like(_UpperCAmelCase ) __snake_case = sequence == pad_token_id __snake_case = 0 return mask def __UpperCAmelCase ( _UpperCAmelCase : Any , _UpperCAmelCase : str , _UpperCAmelCase : int ) -> str: __snake_case = [tokenizer.encode(_UpperCAmelCase ) for line in story_lines] __snake_case = [token for sentence in story_lines_token_ids for token in sentence] __snake_case = [tokenizer.encode(_UpperCAmelCase ) for line in summary_lines] __snake_case = [token for sentence in summary_lines_token_ids for token in sentence] return story_token_ids, summary_token_ids def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] ) -> List[Any]: __snake_case = [] for sequence in batch: __snake_case = -1 __snake_case = [] for s in sequence: if s == separator_token_id: sentence_num += 1 embeddings.append(sentence_num % 2 ) batch_embeddings.append(_UpperCAmelCase ) return torch.tensor(_UpperCAmelCase )
69
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Tuple = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """autoformer""" __SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : List[Any] , a_ : Optional[int] = None , a_ : Optional[int] = None , a_ : str = "student_t" , a_ : str = "nll" , a_ : int = 1 , a_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , a_ : bool = True , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : Optional[List[int]] = None , a_ : Optional[List[int]] = None , a_ : int = 64 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 32 , a_ : int = 32 , a_ : str = "gelu" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : int = 100 , a_ : float = 0.02 , a_ : bool = True , a_ : Union[str, Any]=True , a_ : int = 10 , a_ : int = 25 , a_ : int = 3 , **a_ : Tuple , ): """simple docstring""" __snake_case = prediction_length __snake_case = context_length if context_length is not None else prediction_length __snake_case = distribution_output __snake_case = loss __snake_case = input_size __snake_case = num_time_features __snake_case = lags_sequence __snake_case = scaling __snake_case = num_dynamic_real_features __snake_case = num_static_real_features __snake_case = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case = cardinality else: __snake_case = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case = embedding_dimension else: __snake_case = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case = num_parallel_samples # Transformer architecture configuration __snake_case = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case = d_model __snake_case = encoder_attention_heads __snake_case = decoder_attention_heads __snake_case = encoder_ffn_dim __snake_case = decoder_ffn_dim __snake_case = encoder_layers __snake_case = decoder_layers __snake_case = dropout __snake_case = attention_dropout __snake_case = activation_dropout __snake_case = encoder_layerdrop __snake_case = decoder_layerdrop __snake_case = activation_function __snake_case = init_std __snake_case = use_cache # Autoformer __snake_case = label_length __snake_case = moving_average __snake_case = autocorrelation_factor super().__init__(is_encoder_decoder=a_ , **a_ ) @property def A ( self : Optional[int] ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
69
1
'''simple docstring''' import cmath import math def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> complex: __snake_case = math.radians(_UpperCAmelCase ) __snake_case = math.radians(_UpperCAmelCase ) # Convert voltage and current to rectangular form __snake_case = cmath.rect(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = cmath.rect(_UpperCAmelCase , _UpperCAmelCase ) # Calculate apparent power return voltage_rect * current_rect if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
1
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class SCREAMING_SNAKE_CASE__ : @staticmethod def A ( *a_ : Any , **a_ : List[str] ): """simple docstring""" pass @is_pipeline_test @require_vision @require_timm @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = MODEL_FOR_OBJECT_DETECTION_MAPPING def A ( self : Optional[Any] , a_ : int , a_ : Dict , a_ : List[Any] ): """simple docstring""" __snake_case = ObjectDetectionPipeline(model=a_ , image_processor=a_ ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def A ( self : str , a_ : Dict , a_ : Dict ): """simple docstring""" __snake_case = object_detector("./tests/fixtures/tests_samples/COCO/000000039769.png" , threshold=0.0 ) self.assertGreater(len(a_ ) , 0 ) for detected_object in outputs: self.assertEqual( a_ , { "score": ANY(a_ ), "label": ANY(a_ ), "box": {"xmin": ANY(a_ ), "ymin": ANY(a_ ), "xmax": ANY(a_ ), "ymax": ANY(a_ )}, } , ) import datasets __snake_case = datasets.load_dataset("hf-internal-testing/fixtures_image_utils" , "image" , split="test" ) __snake_case = [ Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ), "http://images.cocodataset.org/val2017/000000039769.jpg", # RGBA dataset[0]["file"], # LA dataset[1]["file"], # L dataset[2]["file"], ] __snake_case = object_detector(a_ , threshold=0.0 ) self.assertEqual(len(a_ ) , len(a_ ) ) for outputs in batch_outputs: self.assertGreater(len(a_ ) , 0 ) for detected_object in outputs: self.assertEqual( a_ , { "score": ANY(a_ ), "label": ANY(a_ ), "box": {"xmin": ANY(a_ ), "ymin": ANY(a_ ), "xmax": ANY(a_ ), "ymax": ANY(a_ )}, } , ) @require_tf @unittest.skip("Object detection not implemented in TF" ) def A ( self : Union[str, Any] ): """simple docstring""" pass @require_torch def A ( self : List[str] ): """simple docstring""" __snake_case = "hf-internal-testing/tiny-detr-mobilenetsv3" __snake_case = AutoModelForObjectDetection.from_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) __snake_case = ObjectDetectionPipeline(model=a_ , feature_extractor=a_ ) __snake_case = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" , threshold=0.0 ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {"score": 0.3376, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, {"score": 0.3376, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, ] , ) __snake_case = object_detector( [ "http://images.cocodataset.org/val2017/000000039769.jpg", "http://images.cocodataset.org/val2017/000000039769.jpg", ] , threshold=0.0 , ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ [ {"score": 0.3376, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, {"score": 0.3376, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, ], [ {"score": 0.3376, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, {"score": 0.3376, "label": "LABEL_0", "box": {"xmin": 159, "ymin": 120, "xmax": 480, "ymax": 359}}, ], ] , ) @require_torch @slow def A ( self : str ): """simple docstring""" __snake_case = "facebook/detr-resnet-50" __snake_case = AutoModelForObjectDetection.from_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) __snake_case = ObjectDetectionPipeline(model=a_ , feature_extractor=a_ ) __snake_case = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {"score": 0.9982, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.9960, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.9955, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.9988, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.9987, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ] , ) __snake_case = object_detector( [ "http://images.cocodataset.org/val2017/000000039769.jpg", "http://images.cocodataset.org/val2017/000000039769.jpg", ] ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ [ {"score": 0.9982, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.9960, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.9955, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.9988, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.9987, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], [ {"score": 0.9982, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.9960, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.9955, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.9988, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.9987, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], ] , ) @require_torch @slow def A ( self : int ): """simple docstring""" __snake_case = "facebook/detr-resnet-50" __snake_case = pipeline("object-detection" , model=a_ ) __snake_case = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {"score": 0.9982, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.9960, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.9955, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.9988, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.9987, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ] , ) __snake_case = object_detector( [ "http://images.cocodataset.org/val2017/000000039769.jpg", "http://images.cocodataset.org/val2017/000000039769.jpg", ] ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ [ {"score": 0.9982, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.9960, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.9955, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.9988, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.9987, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], [ {"score": 0.9982, "label": "remote", "box": {"xmin": 40, "ymin": 70, "xmax": 175, "ymax": 117}}, {"score": 0.9960, "label": "remote", "box": {"xmin": 333, "ymin": 72, "xmax": 368, "ymax": 187}}, {"score": 0.9955, "label": "couch", "box": {"xmin": 0, "ymin": 1, "xmax": 639, "ymax": 473}}, {"score": 0.9988, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.9987, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ], ] , ) @require_torch @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = 0.9985 __snake_case = "facebook/detr-resnet-50" __snake_case = pipeline("object-detection" , model=a_ ) __snake_case = object_detector("http://images.cocodataset.org/val2017/000000039769.jpg" , threshold=a_ ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {"score": 0.9988, "label": "cat", "box": {"xmin": 13, "ymin": 52, "xmax": 314, "ymax": 470}}, {"score": 0.9987, "label": "cat", "box": {"xmin": 345, "ymin": 23, "xmax": 640, "ymax": 368}}, ] , ) @require_torch @require_pytesseract @slow def A ( self : Tuple ): """simple docstring""" __snake_case = "Narsil/layoutlmv3-finetuned-funsd" __snake_case = 0.9993 __snake_case = pipeline("object-detection" , model=a_ , threshold=a_ ) __snake_case = object_detector( "https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png" ) self.assertEqual( nested_simplify(a_ , decimals=4 ) , [ {"score": 0.9993, "label": "I-ANSWER", "box": {"xmin": 294, "ymin": 254, "xmax": 343, "ymax": 264}}, {"score": 0.9993, "label": "I-ANSWER", "box": {"xmin": 294, "ymin": 254, "xmax": 343, "ymax": 264}}, ] , )
69
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 a : Tuple = get_tests_dir('''fixtures''') a : Dict = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : int = get_tests_dir('''fixtures/dummy-config.json''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Tuple ): """simple docstring""" __snake_case = 0 def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __snake_case = AutoFeatureExtractor.from_pretrained(a_ ).to_dict() config_dict.pop("feature_extractor_type" ) __snake_case = WavaVecaFeatureExtractor(**a_ ) # save in new folder model_config.save_pretrained(a_ ) config.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) # make sure private variable is not incorrectly saved __snake_case = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with self.assertRaisesRegex( a_ , "bert-base is not a local folder and is not a valid model identifier" ): __snake_case = AutoFeatureExtractor.from_pretrained("bert-base" ) def A ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( a_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __snake_case = AutoFeatureExtractor.from_pretrained(a_ , revision="aaaaaa" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( a_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): __snake_case = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ , trust_remote_code=a_ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def A ( self : int ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoFeatureExtractor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def A ( self : Dict ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # If remote code is not set, the default is to use local __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(a_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
69
1
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging a : List[Any] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = ["""input_values""", """padding_mask"""] def __init__( self : Tuple , a_ : int = 1 , a_ : int = 24_000 , a_ : float = 0.0 , a_ : float = None , a_ : float = None , **a_ : Dict , ): """simple docstring""" super().__init__(feature_size=a_ , sampling_rate=a_ , padding_value=a_ , **a_ ) __snake_case = chunk_length_s __snake_case = overlap @property def A ( self : str ): """simple docstring""" if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def A ( self : Optional[Any] ): """simple docstring""" if self.chunk_length_s is None or self.overlap is None: return None else: return max(1 , int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self : List[str] , a_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a_ : Optional[Union[bool, str, PaddingStrategy]] = None , a_ : Optional[bool] = False , a_ : Optional[int] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : Optional[int] = None , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if padding and truncation: raise ValueError("Both padding and truncation were set. Make sure you only set one." ) elif padding is None: # by default let's pad the inputs __snake_case = True __snake_case = bool( isinstance(a_ , (list, tuple) ) and (isinstance(raw_audio[0] , (np.ndarray, tuple, list) )) ) if is_batched: __snake_case = [np.asarray(a_ , dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(a_ , np.ndarray ): __snake_case = np.asarray(a_ , dtype=np.floataa ) elif isinstance(a_ , np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): __snake_case = raw_audio.astype(np.floataa ) # always return batch if not is_batched: __snake_case = [np.asarray(a_ ).T] # verify inputs are valid for idx, example in enumerate(a_ ): if example.ndim > 2: raise ValueError(f'''Expected input shape (channels, length) but got shape {example.shape}''' ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(f'''Expected mono audio but example has {example.shape[-1]} channels''' ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(f'''Expected stereo audio but example has {example.shape[-1]} channels''' ) __snake_case = None __snake_case = BatchFeature({"input_values": raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: __snake_case = min(array.shape[0] for array in raw_audio ) __snake_case = int(np.floor(max_length / self.chunk_stride ) ) __snake_case = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: __snake_case = max(array.shape[0] for array in raw_audio ) __snake_case = int(np.ceil(max_length / self.chunk_stride ) ) __snake_case = (nb_step - 1) * self.chunk_stride + self.chunk_length __snake_case = "max_length" else: __snake_case = input_values # normal padding on batch if padded_inputs is None: __snake_case = self.pad( a_ , max_length=a_ , truncation=a_ , padding=a_ , return_attention_mask=a_ , ) if padding: __snake_case = padded_inputs.pop("attention_mask" ) __snake_case = [] for example in padded_inputs.pop("input_values" ): if self.feature_size == 1: __snake_case = example[..., None] input_values.append(example.T ) __snake_case = input_values if return_tensors is not None: __snake_case = padded_inputs.convert_to_tensors(a_ ) return padded_inputs
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise TypeError("Input value must be an 'int' type" ) __snake_case = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : list[int] ) -> list[list[int]]: __snake_case = [] if len(_UpperCAmelCase ) == 1: return [nums.copy()] for _ in range(len(_UpperCAmelCase ) ): __snake_case = nums.pop(0 ) __snake_case = permute(_UpperCAmelCase ) for perm in permutations: perm.append(_UpperCAmelCase ) result.extend(_UpperCAmelCase ) nums.append(_UpperCAmelCase ) return result def __UpperCAmelCase ( _UpperCAmelCase : str ) -> List[Any]: def backtrack(_UpperCAmelCase : Union[str, Any] ): if start == len(_UpperCAmelCase ) - 1: output.append(nums[:] ) else: for i in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case , __snake_case = nums[i], nums[start] backtrack(start + 1 ) __snake_case , __snake_case = nums[i], nums[start] # backtrack __snake_case = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function a : Union[str, Any] = permutea([1, 2, 3]) print(res) doctest.testmod()
69
'''simple docstring''' a : Dict = range(2, 20 + 1) a : Optional[int] = [10**k for k in range(ks[-1] + 1)] a : dict[int, dict[int, list[list[int]]]] = {} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] ) -> int: __snake_case = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) __snake_case = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) __snake_case , __snake_case = 0, 0 __snake_case = n - i __snake_case = memo.get(_UpperCAmelCase ) if sub_memo is not None: __snake_case = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over __snake_case = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __snake_case = _k break if max_jump >= 0: __snake_case , __snake_case , __snake_case = jumps[max_jump] # since the difference between jumps is cached, add c __snake_case = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: __snake_case = [] else: __snake_case = {c: []} __snake_case = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __snake_case , __snake_case = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __snake_case , __snake_case = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped __snake_case = sub_memo[c] # keep jumps sorted by # of terms skipped __snake_case = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __snake_case = i __snake_case , __snake_case , __snake_case = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __snake_case = ds_c + ds_b diff += addend __snake_case = 0 for j in range(_UpperCAmelCase ): __snake_case = a_i[j] + addend __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Tuple: for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case = digits[j] + addend if s >= 10: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) __snake_case = addend // 10 + quotient else: __snake_case = s __snake_case = addend // 10 if addend == 0: break while addend > 0: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int = 10**15 ) -> int: __snake_case = [1] __snake_case = 1 __snake_case = 0 while True: __snake_case , __snake_case = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break __snake_case = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
69
1
'''simple docstring''' import numpy as np import torch from torch.utils.data import Dataset from utils import logger class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Optional[int] , a_ : int , a_ : str ): """simple docstring""" __snake_case = params __snake_case = np.array(a_ ) __snake_case = np.array([len(a_ ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : str , a_ : Any ): """simple docstring""" return (self.token_ids[index], self.lengths[index]) def __len__( self : int ): """simple docstring""" return len(self.lengths ) def A ( self : Tuple ): """simple docstring""" assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def A ( self : Any ): """simple docstring""" __snake_case = self.params.max_model_input_size __snake_case = self.lengths > max_len logger.info(f'''Splitting {sum(a_ )} too long sequences.''' ) def divide_chunks(a_ : Optional[int] , a_ : int ): return [l[i : i + n] for i in range(0 , len(a_ ) , a_ )] __snake_case = [] __snake_case = [] if self.params.mlm: __snake_case , __snake_case = self.params.special_tok_ids["cls_token"], self.params.special_tok_ids["sep_token"] else: __snake_case , __snake_case = self.params.special_tok_ids["bos_token"], self.params.special_tok_ids["eos_token"] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: __snake_case = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: __snake_case = np.insert(a_ , 0 , a_ ) if sub_s[-1] != sep_id: __snake_case = np.insert(a_ , len(a_ ) , a_ ) assert len(a_ ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(a_ ) new_tok_ids.extend(a_ ) new_lengths.extend([len(a_ ) for l in sub_seqs] ) __snake_case = np.array(a_ ) __snake_case = np.array(a_ ) def A ( self : Any ): """simple docstring""" __snake_case = len(self ) __snake_case = self.lengths > 11 __snake_case = self.token_ids[indices] __snake_case = self.lengths[indices] __snake_case = len(self ) logger.info(f'''Remove {init_size - new_size} too short (<=11 tokens) sequences.''' ) def A ( self : List[Any] ): """simple docstring""" if "unk_token" not in self.params.special_tok_ids: return else: __snake_case = self.params.special_tok_ids["unk_token"] __snake_case = len(self ) __snake_case = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) __snake_case = (unk_occs / self.lengths) < 0.5 __snake_case = self.token_ids[indices] __snake_case = self.lengths[indices] __snake_case = len(self ) logger.info(f'''Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).''' ) def A ( self : str ): """simple docstring""" if not self.params.is_master: return logger.info(f'''{len(self )} sequences''' ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def A ( self : List[str] , a_ : Union[str, Any] ): """simple docstring""" __snake_case = [t[0] for t in batch] __snake_case = [t[1] for t in batch] assert len(a_ ) == len(a_ ) # Max for paddings __snake_case = max(a_ ) # Pad token ids if self.params.mlm: __snake_case = self.params.special_tok_ids["pad_token"] else: __snake_case = self.params.special_tok_ids["unk_token"] __snake_case = [list(t.astype(a_ ) ) + [pad_idx] * (max_seq_len_ - len(a_ )) for t in token_ids] assert len(tk_ ) == len(a_ ) assert all(len(a_ ) == max_seq_len_ for t in tk_ ) __snake_case = torch.tensor(tk_ ) # (bs, max_seq_len_) __snake_case = torch.tensor(a_ ) # (bs) return tk_t, lg_t
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
1
'''simple docstring''' import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class SCREAMING_SNAKE_CASE__ : def __init__( self : List[Any] , a_ : Optional[Any]=2 , a_ : Tuple=3 , a_ : Any=64 , a_ : str=None ): """simple docstring""" __snake_case = np.random.default_rng(a_ ) __snake_case = length __snake_case = rng.normal(size=(length,) ).astype(np.floataa ) __snake_case = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self : Optional[Any] ): """simple docstring""" return self.length def __getitem__( self : Optional[Any] , a_ : List[str] ): """simple docstring""" return {"x": self.x[i], "y": self.y[i]} class SCREAMING_SNAKE_CASE__ ( torch.nn.Module ): def __init__( self : Any , a_ : List[str]=0 , a_ : Union[str, Any]=0 , a_ : Tuple=False ): """simple docstring""" super().__init__() __snake_case = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __snake_case = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) __snake_case = True def A ( self : List[str] , a_ : int=None ): """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) __snake_case = False return x * self.a[0] + self.b[0] class SCREAMING_SNAKE_CASE__ ( torch.nn.Module ): def __init__( self : Dict , a_ : Dict=0 , a_ : Any=0 , a_ : Optional[Any]=False ): """simple docstring""" super().__init__() __snake_case = torch.nn.Parameter(torch.tensor(a_ ).float() ) __snake_case = torch.nn.Parameter(torch.tensor(a_ ).float() ) __snake_case = True def A ( self : Dict , a_ : Any=None ): """simple docstring""" if self.first_batch: print(f'''Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}''' ) __snake_case = False return x * self.a + self.b def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : int = 16 ) -> str: from datasets import load_dataset from transformers import AutoTokenizer __snake_case = AutoTokenizer.from_pretrained("bert-base-cased" ) __snake_case = {"train": "tests/test_samples/MRPC/train.csv", "validation": "tests/test_samples/MRPC/dev.csv"} __snake_case = load_dataset("csv" , data_files=_UpperCAmelCase ) __snake_case = datasets["train"].unique("label" ) __snake_case = {v: i for i, v in enumerate(_UpperCAmelCase )} def tokenize_function(_UpperCAmelCase : List[str] ): # max_length=None => use the model max length (it's actually the default) __snake_case = tokenizer( examples["sentence1"] , examples["sentence2"] , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" ) if "label" in examples: __snake_case = [label_to_id[l] for l in examples["label"]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __snake_case = datasets.map( _UpperCAmelCase , batched=_UpperCAmelCase , remove_columns=["sentence1", "sentence2", "label"] , ) def collate_fn(_UpperCAmelCase : Optional[Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(_UpperCAmelCase , padding="max_length" , max_length=1_28 , return_tensors="pt" ) return tokenizer.pad(_UpperCAmelCase , padding="longest" , return_tensors="pt" ) # Instantiate dataloaders. __snake_case = DataLoader(tokenized_datasets["train"] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=2 ) __snake_case = DataLoader(tokenized_datasets["validation"] , shuffle=_UpperCAmelCase , collate_fn=_UpperCAmelCase , batch_size=1 ) return train_dataloader, eval_dataloader
69
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
1
'''simple docstring''' from typing import Dict from transformers import EvalPrediction, HfArgumentParser, TrainingArguments, is_torch_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, get_torch_dist_unique_port, require_torch_multi_gpu, require_torch_neuroncore, ) from transformers.training_args import ParallelMode from transformers.utils import logging a : Union[str, Any] = logging.get_logger(__name__) if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset from transformers import Trainer class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Union[str, Any] , a_ : int = 101 ): """simple docstring""" __snake_case = length def __len__( self : Optional[Any] ): """simple docstring""" return self.length def __getitem__( self : Optional[Any] , a_ : Any ): """simple docstring""" return i class SCREAMING_SNAKE_CASE__ : def __call__( self : List[Any] , a_ : Any ): """simple docstring""" return {"input_ids": torch.tensor(a_ ), "labels": torch.tensor(a_ )} class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : Union[str, Any] ): """simple docstring""" super().__init__() # Add some (unused) params otherwise DDP will complain. __snake_case = nn.Linear(120 , 80 ) def A ( self : int , a_ : Optional[int] , a_ : List[Any]=None ): """simple docstring""" if labels is not None: return torch.tensor(0.0 , device=input_ids.device ), input_ids else: return input_ids class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @require_torch_neuroncore def A ( self : List[Any] ): """simple docstring""" __snake_case = f'''--nproc_per_node=2 --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py '''.split() __snake_case = self.get_auto_remove_tmp_dir() __snake_case = f'''--output_dir {output_dir}'''.split() __snake_case = ["torchrun"] + distributed_args + args execute_subprocess_async(a_ , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @require_torch_multi_gpu def A ( self : Any ): """simple docstring""" __snake_case = f'''--nproc_per_node={torch.cuda.device_count()} --master_port={get_torch_dist_unique_port()} {self.test_file_dir}/test_trainer_distributed.py '''.split() __snake_case = self.get_auto_remove_tmp_dir() __snake_case = f'''--output_dir {output_dir}'''.split() __snake_case = ["torchrun"] + distributed_args + args execute_subprocess_async(a_ , env=self.get_env() ) # successful return here == success - any errors would have caused an error in the sub-call if __name__ == "__main__": # The script below is meant to be run under torch.distributed, on a machine with multiple GPUs: # # PYTHONPATH="src" python -m torch.distributed.run --nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py a : Optional[Any] = HfArgumentParser((TrainingArguments,)) a : List[str] = parser.parse_args_into_dataclasses()[0] logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}, ''' F'''distributed training: {training_args.parallel_mode != ParallelMode.NOT_DISTRIBUTED}''' ) # Essentially, what we want to verify in the distributed case is that we get all samples back, # in the right order. (this is crucial for prediction for instance) for dataset_length in [101, 40, 7]: a : Union[str, Any] = DummyDataset(dataset_length) def __UpperCAmelCase ( _UpperCAmelCase : EvalPrediction ) -> Dict: __snake_case = list(range(len(_UpperCAmelCase ) ) ) __snake_case = p.predictions.tolist() == sequential and p.label_ids.tolist() == sequential if not success and training_args.local_rank == 0: logger.warning( "Predictions and/or labels do not match expected results:\n - predictions: " F'''{p.predictions.tolist()}\n - labels: {p.label_ids.tolist()}\n - expected: {sequential}''' ) return {"success": success} a : List[Any] = Trainer( model=DummyModel(), args=training_args, data_collator=DummyDataCollator(), eval_dataset=dataset, compute_metrics=compute_metrics, ) a : Optional[Any] = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) a : Tuple = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) a : Any = 2 a : Union[str, Any] = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) a : int = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) a : str = None
69
'''simple docstring''' import mpmath # for roots of unity import numpy as np class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , a_ : Optional[int]=None , a_ : int=None ): """simple docstring""" __snake_case = list(poly_a or [0] )[:] __snake_case = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() __snake_case = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() __snake_case = len(self.polyB ) # Add 0 to make lengths equal a power of 2 __snake_case = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform __snake_case = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product __snake_case = self.__multiply() def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" __snake_case = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB] # Corner case if len(a_ ) <= 1: return dft[0] # __snake_case = self.c_max_length // 2 while next_ncol > 0: __snake_case = [[] for i in range(a_ )] __snake_case = self.root**next_ncol # First half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update __snake_case = new_dft __snake_case = next_ncol // 2 return dft[0] def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.__dft("A" ) __snake_case = self.__dft("B" ) __snake_case = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT __snake_case = 2 while next_ncol <= self.c_max_length: __snake_case = [[] for i in range(a_ )] __snake_case = self.root ** (next_ncol // 2) __snake_case = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update __snake_case = new_inverse_c next_ncol *= 2 # Unpack __snake_case = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Optional[int] ): """simple docstring""" __snake_case = "A = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) __snake_case = "B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) __snake_case = "A*B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import inspect import unittest from transformers import ViTMSNConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMSNForImageClassification, ViTMSNModel from transformers.models.vit_msn.modeling_vit_msn import VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class SCREAMING_SNAKE_CASE__ : def __init__( self : List[str] , a_ : str , a_ : Any=13 , a_ : Union[str, Any]=30 , a_ : Union[str, Any]=2 , a_ : Union[str, Any]=3 , a_ : List[str]=True , a_ : Union[str, Any]=True , a_ : Union[str, Any]=32 , a_ : List[str]=5 , a_ : Union[str, Any]=4 , a_ : Union[str, Any]=37 , a_ : Tuple="gelu" , a_ : str=0.1 , a_ : Optional[Any]=0.1 , a_ : Optional[Any]=10 , a_ : Tuple=0.02 , a_ : Union[str, Any]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = image_size __snake_case = patch_size __snake_case = num_channels __snake_case = is_training __snake_case = use_labels __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = scope # in ViT MSN, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __snake_case = (image_size // patch_size) ** 2 __snake_case = num_patches + 1 def A ( self : List[str] ): """simple docstring""" __snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = self.get_config() return config, pixel_values, labels def A ( self : int ): """simple docstring""" return ViTMSNConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , initializer_range=self.initializer_range , ) def A ( self : str , a_ : Tuple , a_ : Optional[int] , a_ : int ): """simple docstring""" __snake_case = ViTMSNModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Dict , a_ : Union[str, Any] , a_ : Any , a_ : List[Any] ): """simple docstring""" __snake_case = self.type_sequence_label_size __snake_case = ViTMSNForImageClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , labels=a_ ) print("Pixel and labels shape: {pixel_values.shape}, {labels.shape}" ) print("Labels: {labels}" ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __snake_case = 1 __snake_case = ViTMSNForImageClassification(a_ ) model.to(a_ ) model.eval() __snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __snake_case = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def A ( self : Tuple ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case = config_and_inputs __snake_case = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = (ViTMSNModel, ViTMSNForImageClassification) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( {"""feature-extraction""": ViTMSNModel, """image-classification""": ViTMSNForImageClassification} if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[int] ): """simple docstring""" __snake_case = ViTMSNModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , has_text_modality=a_ , hidden_size=37 ) def A ( self : Any ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="ViTMSN does not use inputs_embeds" ) def A ( self : Any ): """simple docstring""" pass def A ( self : str ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = model_class(a_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ , nn.Linear ) ) def A ( self : Dict ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = model_class(a_ ) __snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case = [*signature.parameters.keys()] __snake_case = ["pixel_values"] self.assertListEqual(arg_names[:1] , a_ ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) @slow def A ( self : int ): """simple docstring""" for model_name in VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = ViTMSNModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def __UpperCAmelCase ( ) -> Any: __snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @cached_property def A ( self : List[Any] ): """simple docstring""" return ViTImageProcessor.from_pretrained("facebook/vit-msn-small" ) if is_vision_available() else None @slow def A ( self : List[str] ): """simple docstring""" torch.manual_seed(2 ) __snake_case = ViTMSNForImageClassification.from_pretrained("facebook/vit-msn-small" ).to(a_ ) __snake_case = self.default_image_processor __snake_case = prepare_img() __snake_case = image_processor(images=a_ , return_tensors="pt" ).to(a_ ) # forward pass with torch.no_grad(): __snake_case = model(**a_ ) # verify the logits __snake_case = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a_ ) __snake_case = torch.tensor([-0.0803, -0.4454, -0.2375] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a_ , atol=1e-4 ) )
69
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
1
'''simple docstring''' import tensorflow as tf from ...tf_utils import shape_list class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): def __init__( self : str , a_ : str , a_ : Optional[int] , a_ : Optional[Any] , a_ : List[str] , a_ : int=1 , a_ : str=False , **a_ : Optional[int] ): """simple docstring""" super().__init__(**a_ ) __snake_case = vocab_size __snake_case = d_embed __snake_case = d_proj __snake_case = cutoffs + [vocab_size] __snake_case = [0] + self.cutoffs __snake_case = div_val __snake_case = self.cutoffs[0] __snake_case = len(self.cutoffs ) - 1 __snake_case = self.shortlist_size + self.n_clusters __snake_case = keep_order __snake_case = [] __snake_case = [] def A ( self : int , a_ : List[str] ): """simple docstring""" if self.n_clusters > 0: __snake_case = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer="zeros" , trainable=a_ , name="cluster_weight" ) __snake_case = self.add_weight( shape=(self.n_clusters,) , initializer="zeros" , trainable=a_ , name="cluster_bias" ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: __snake_case = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer="zeros" , trainable=a_ , name=f'''out_projs_._{i}''' , ) self.out_projs.append(a_ ) else: self.out_projs.append(a_ ) __snake_case = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer="zeros" , trainable=a_ , name=f'''out_layers_._{i}_._weight''' , ) __snake_case = self.add_weight( shape=(self.vocab_size,) , initializer="zeros" , trainable=a_ , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): __snake_case , __snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] __snake_case = self.d_embed // (self.div_val**i) __snake_case = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer="zeros" , trainable=a_ , name=f'''out_projs_._{i}''' ) self.out_projs.append(a_ ) __snake_case = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer="zeros" , trainable=a_ , name=f'''out_layers_._{i}_._weight''' , ) __snake_case = self.add_weight( shape=(r_idx - l_idx,) , initializer="zeros" , trainable=a_ , name=f'''out_layers_._{i}_._bias''' , ) self.out_layers.append((weight, bias) ) super().build(a_ ) @staticmethod def A ( a_ : str , a_ : Tuple , a_ : Tuple , a_ : List[str]=None ): """simple docstring""" __snake_case = x if proj is not None: __snake_case = tf.einsum("ibd,ed->ibe" , a_ , a_ ) return tf.einsum("ibd,nd->ibn" , a_ , a_ ) + b @staticmethod def A ( a_ : int , a_ : Optional[Any] ): """simple docstring""" __snake_case = shape_list(a_ ) __snake_case = tf.range(lp_size[0] , dtype=target.dtype ) __snake_case = tf.stack([r, target] , 1 ) return tf.gather_nd(a_ , a_ ) def A ( self : Optional[int] , a_ : Dict , a_ : Dict , a_ : Tuple=True , a_ : Optional[Any]=False ): """simple docstring""" __snake_case = 0 if self.n_clusters == 0: __snake_case = self._logit(a_ , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: __snake_case = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=a_ , logits=a_ ) __snake_case = tf.nn.log_softmax(a_ , axis=-1 ) else: __snake_case = shape_list(a_ ) __snake_case = [] __snake_case = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): __snake_case , __snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: __snake_case = (target >= l_idx) & (target < r_idx) __snake_case = tf.where(a_ ) __snake_case = tf.boolean_mask(a_ , a_ ) - l_idx if self.div_val == 1: __snake_case = self.out_layers[0][0][l_idx:r_idx] __snake_case = self.out_layers[0][1][l_idx:r_idx] else: __snake_case = self.out_layers[i][0] __snake_case = self.out_layers[i][1] if i == 0: __snake_case = tf.concat([cur_W, self.cluster_weight] , 0 ) __snake_case = tf.concat([cur_b, self.cluster_bias] , 0 ) __snake_case = self._logit(a_ , a_ , a_ , self.out_projs[0] ) __snake_case = tf.nn.log_softmax(a_ ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: __snake_case = tf.boolean_mask(a_ , a_ ) __snake_case = self._gather_logprob(a_ , a_ ) else: __snake_case = self._logit(a_ , a_ , a_ , self.out_projs[i] ) __snake_case = tf.nn.log_softmax(a_ ) __snake_case = self.cutoffs[0] + i - 1 # No probability for the head cluster __snake_case = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(a_ ) if target is not None: __snake_case = tf.boolean_mask(a_ , a_ ) __snake_case = tf.boolean_mask(a_ , a_ ) __snake_case = self._gather_logprob(a_ , a_ ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(a_ , -cur_logprob , shape_list(a_ ) ) __snake_case = tf.concat(a_ , axis=-1 ) if target is not None: if return_mean: __snake_case = tf.reduce_mean(a_ ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(a_ ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(a_ , name=self.name , aggregation="mean" if return_mean else "" ) return out
69
'''simple docstring''' import json import os import torch from diffusers import UNetaDModel os.makedirs('''hub/hopper-medium-v2/unet/hor32''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/unet/hor128''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/value_function''', exist_ok=True) def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> str: if hor == 1_28: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D") elif hor == 32: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 64, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D", "UpResnetBlock1D") __snake_case = torch.load(F'''/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch''' ) __snake_case = model.state_dict() __snake_case = { "down_block_types": down_block_types, "block_out_channels": block_out_channels, "up_block_types": up_block_types, "layers_per_block": 1, "use_timestep_embedding": True, "out_block_type": "OutConv1DBlock", "norm_num_groups": 8, "downsample_each_block": False, "in_channels": 14, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "flip_sin_to_cos": False, "freq_shift": 1, "sample_size": 6_55_36, "mid_block_type": "MidResTemporalBlock1D", "act_fn": "mish", } __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , F'''hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin''' ) with open(F'''hub/hopper-medium-v2/unet/hor{hor}/config.json''' , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) def __UpperCAmelCase ( ) -> List[Any]: __snake_case = { "in_channels": 14, "down_block_types": ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D"), "up_block_types": (), "out_block_type": "ValueFunction", "mid_block_type": "ValueFunctionMidBlock1D", "block_out_channels": (32, 64, 1_28, 2_56), "layers_per_block": 1, "downsample_each_block": True, "sample_size": 6_55_36, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "use_timestep_embedding": True, "flip_sin_to_cos": False, "freq_shift": 1, "norm_num_groups": 8, "act_fn": "mish", } __snake_case = torch.load("/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch" ) __snake_case = model __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , "hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin" ) with open("hub/hopper-medium-v2/value_function/config.json" , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": unet(32) # unet(128) value_function()
69
1
'''simple docstring''' import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : int = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = AlbertTokenizer __SCREAMING_SNAKE_CASE = AlbertTokenizerFast __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = True def A ( self : Optional[Any] ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = AlbertTokenizer(a_ ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : Union[str, Any] , a_ : Optional[int] ): """simple docstring""" __snake_case = "this is a test" __snake_case = "this is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<pad>" __snake_case = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<pad>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "▁eloquent" ) self.assertEqual(len(a_ ) , 30_000 ) def A ( self : Any ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 30_000 ) def A ( self : str ): """simple docstring""" if not self.test_rust_tokenizer: return __snake_case = self.get_tokenizer() __snake_case = self.get_rust_tokenizer() __snake_case = "I was born in 92000, and this is falsé." __snake_case = tokenizer.tokenize(a_ ) __snake_case = rust_tokenizer.tokenize(a_ ) self.assertListEqual(a_ , a_ ) __snake_case = tokenizer.encode(a_ , add_special_tokens=a_ ) __snake_case = rust_tokenizer.encode(a_ , add_special_tokens=a_ ) self.assertListEqual(a_ , a_ ) __snake_case = self.get_rust_tokenizer() __snake_case = tokenizer.encode(a_ ) __snake_case = rust_tokenizer.encode(a_ ) self.assertListEqual(a_ , a_ ) def A ( self : int ): """simple docstring""" __snake_case = AlbertTokenizer(a_ , keep_accents=a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁this", "▁is", "▁a", "▁test"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [48, 25, 21, 1_289] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( a_ , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", "."] ) __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual(a_ , [31, 23, 386, 19, 561, 3_050, 15, 17, 48, 25, 8_256, 18, 1, 9] ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) self.assertListEqual( a_ , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "."] , ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = AlbertTokenizer(a_ ) __snake_case = tokenizer.encode("sequence builders" ) __snake_case = tokenizer.encode("multi-sequence build" ) __snake_case = tokenizer.build_inputs_with_special_tokens(a_ ) __snake_case = tokenizer.build_inputs_with_special_tokens(a_ , a_ ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = {"attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "input_ids": [[2, 21_970, 13, 5, 6_092, 167, 28, 7_103, 2_153, 673, 8, 7_028, 12_051, 18, 17, 7_103, 2_153, 673, 8, 3_515, 18_684, 8, 4_461, 6, 1_927, 297, 8, 12_060, 2_607, 18, 13, 5, 4_461, 15, 10_538, 38, 8, 135, 15, 822, 58, 15, 993, 10_363, 15, 1_460, 8_005, 4_461, 15, 993, 255, 2_328, 9, 9, 9, 6, 26, 1_112, 816, 3_260, 13, 5, 103, 2_377, 6, 17, 1_112, 816, 2_782, 13, 5, 103, 10_641, 6, 29, 84, 2_512, 2_430, 782, 18_684, 2_761, 19, 808, 2_430, 2_556, 17, 855, 1_480, 9_477, 4_091, 128, 11_712, 15, 7_103, 2_153, 673, 17, 24_883, 9_990, 9, 3], [2, 11_502, 25, 1_006, 20, 782, 8, 11_809, 855, 1_732, 19_393, 18_667, 37, 367, 21_018, 69, 1_854, 34, 11_860, 19_124, 27, 156, 225, 17, 193, 4_141, 19, 65, 9_124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2_231, 886, 2_385, 17_659, 84, 14, 16_792, 1_952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="albert-base-v2" , revision="6b6560eaf5ff2e250b00c50f380c5389a9c2d82e" , )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 1_00_00_00 ) -> int: __snake_case = 1 __snake_case = 1 __snake_case = {1: 1} for inputa in range(2 , _UpperCAmelCase ): __snake_case = 0 __snake_case = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __snake_case = (3 * number) + 1 counter += 1 if inputa not in counters: __snake_case = counter if counter > pre_counter: __snake_case = inputa __snake_case = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
69
1
'''simple docstring''' import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) a : Tuple = logging.getLogger(__name__) @dataclass(frozen=_UpperCamelCase ) class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None @dataclass(frozen=_UpperCamelCase ) class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None if is_torch_available(): import torch from torch.utils.data import Dataset class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = 42 def __init__( self : List[Any] , a_ : str , a_ : PreTrainedTokenizer , a_ : str , a_ : Optional[int] = None , a_ : List[Any]=False , a_ : bool = False , ): """simple docstring""" __snake_case = hans_processors[task]() __snake_case = os.path.join( a_ , "cached_{}_{}_{}_{}".format( "dev" if evaluate else "train" , tokenizer.__class__.__name__ , str(a_ ) , a_ , ) , ) __snake_case = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) __snake_case , __snake_case = label_list[2], label_list[1] __snake_case = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __snake_case = cached_features_file + ".lock" with FileLock(a_ ): if os.path.exists(a_ ) and not overwrite_cache: logger.info(f'''Loading features from cached file {cached_features_file}''' ) __snake_case = torch.load(a_ ) else: logger.info(f'''Creating features from dataset file at {data_dir}''' ) __snake_case = ( processor.get_dev_examples(a_ ) if evaluate else processor.get_train_examples(a_ ) ) logger.info("Training examples: %s" , len(a_ ) ) __snake_case = hans_convert_examples_to_features(a_ , a_ , a_ , a_ ) logger.info("Saving features into cached file %s" , a_ ) torch.save(self.features , a_ ) def __len__( self : Any ): """simple docstring""" return len(self.features ) def __getitem__( self : int , a_ : List[str] ): """simple docstring""" return self.features[i] def A ( self : Dict ): """simple docstring""" return self.label_list if is_tf_available(): import tensorflow as tf class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = 42 def __init__( self : Tuple , a_ : str , a_ : PreTrainedTokenizer , a_ : str , a_ : Optional[int] = 128 , a_ : List[Any]=False , a_ : bool = False , ): """simple docstring""" __snake_case = hans_processors[task]() __snake_case = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) __snake_case , __snake_case = label_list[2], label_list[1] __snake_case = label_list __snake_case = processor.get_dev_examples(a_ ) if evaluate else processor.get_train_examples(a_ ) __snake_case = hans_convert_examples_to_features(a_ , a_ , a_ , a_ ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc="convert examples to features" ): if ex_index % 10_000 == 0: logger.info("Writing example %d of %d" % (ex_index, len(a_ )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) __snake_case = tf.data.Dataset.from_generator( a_ , ( { "example_id": tf.intaa, "input_ids": tf.intaa, "attention_mask": tf.intaa, "token_type_ids": tf.intaa, }, tf.intaa, ) , ( { "example_id": tf.TensorShape([] ), "input_ids": tf.TensorShape([None, None] ), "attention_mask": tf.TensorShape([None, None] ), "token_type_ids": tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def A ( self : List[str] ): """simple docstring""" return self.dataset def __len__( self : Optional[Any] ): """simple docstring""" return len(self.features ) def __getitem__( self : Tuple , a_ : Tuple ): """simple docstring""" return self.features[i] def A ( self : int ): """simple docstring""" return self.label_list class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : List[str] , a_ : int ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(a_ , "heuristics_train_set.txt" ) ) , "train" ) def A ( self : List[str] , a_ : Optional[Any] ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(a_ , "heuristics_evaluation_set.txt" ) ) , "dev" ) def A ( self : str ): """simple docstring""" return ["contradiction", "entailment", "neutral"] def A ( self : Tuple , a_ : int , a_ : int ): """simple docstring""" __snake_case = [] for i, line in enumerate(a_ ): if i == 0: continue __snake_case = "%s-%s" % (set_type, line[0]) __snake_case = line[5] __snake_case = line[6] __snake_case = line[7][2:] if line[7].startswith("ex" ) else line[7] __snake_case = line[0] examples.append(InputExample(guid=a_ , text_a=a_ , text_b=a_ , label=a_ , pairID=a_ ) ) return examples def __UpperCAmelCase ( _UpperCAmelCase : List[InputExample] , _UpperCAmelCase : List[str] , _UpperCAmelCase : int , _UpperCAmelCase : PreTrainedTokenizer , ) -> Optional[Any]: __snake_case = {label: i for i, label in enumerate(_UpperCAmelCase )} __snake_case = [] for ex_index, example in tqdm.tqdm(enumerate(_UpperCAmelCase ) , desc="convert examples to features" ): if ex_index % 1_00_00 == 0: logger.info("Writing example %d" % (ex_index) ) __snake_case = tokenizer( example.text_a , example.text_b , add_special_tokens=_UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" , truncation=_UpperCAmelCase , return_overflowing_tokens=_UpperCAmelCase , ) __snake_case = label_map[example.label] if example.label in label_map else 0 __snake_case = int(example.pairID ) features.append(InputFeatures(**_UpperCAmelCase , label=_UpperCAmelCase , pairID=_UpperCAmelCase ) ) for i, example in enumerate(examples[:5] ): logger.info("*** Example ***" ) logger.info(F'''guid: {example}''' ) logger.info(F'''features: {features[i]}''' ) return features a : Optional[Any] = { '''hans''': 3, } a : str = { '''hans''': HansProcessor, }
69
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """SpeechT5FeatureExtractor""" __SCREAMING_SNAKE_CASE = """SpeechT5Tokenizer""" def __init__( self : List[Any] , a_ : str , a_ : str ): """simple docstring""" super().__init__(a_ , a_ ) def __call__( self : Dict , *a_ : Tuple , **a_ : List[str] ): """simple docstring""" __snake_case = kwargs.pop("audio" , a_ ) __snake_case = kwargs.pop("text" , a_ ) __snake_case = kwargs.pop("text_target" , a_ ) __snake_case = kwargs.pop("audio_target" , a_ ) __snake_case = kwargs.pop("sampling_rate" , a_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __snake_case = self.feature_extractor(a_ , *a_ , sampling_rate=a_ , **a_ ) elif text is not None: __snake_case = self.tokenizer(a_ , **a_ ) else: __snake_case = None if audio_target is not None: __snake_case = self.feature_extractor(audio_target=a_ , *a_ , sampling_rate=a_ , **a_ ) __snake_case = targets["input_values"] elif text_target is not None: __snake_case = self.tokenizer(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : str , **a_ : Dict ): """simple docstring""" __snake_case = kwargs.pop("input_values" , a_ ) __snake_case = kwargs.pop("input_ids" , a_ ) __snake_case = kwargs.pop("labels" , a_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) elif input_ids is not None: __snake_case = self.tokenizer.pad(a_ , **a_ ) else: __snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(a_ , a_ ) and "input_ids" in labels[0]): __snake_case = self.tokenizer.pad(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = self.feature_extractor.feature_size __snake_case = self.feature_extractor.num_mel_bins __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) __snake_case = feature_size_hack __snake_case = targets["input_values"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : Any , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : Optional[int] , *a_ : Union[str, Any] , **a_ : str ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ )
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : int ) -> float: if principal <= 0: raise Exception("Principal borrowed must be > 0" ) if rate_per_annum < 0: raise Exception("Rate of interest must be >= 0" ) if years_to_repay <= 0 or not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise Exception("Years to repay must be an integer > 0" ) # Yearly rate is divided by 12 to get monthly rate __snake_case = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly __snake_case = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : Optional[Any] , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __snake_case = input_file.read() __snake_case = regexp.search(a_ ) return match def A ( self : Any , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __snake_case = regexp.finditer(a_ ) __snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A ( self : Optional[int] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a_ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(a_ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
1
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging a : Optional[Any] = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = ["""input_features""", """attention_mask"""] def __init__( self : Tuple , a_ : Dict=80 , a_ : Tuple=16_000 , a_ : Optional[Any]=0.0 , a_ : Optional[int]=10 , a_ : List[Any]=25 , a_ : List[str]="hamming_window" , a_ : int=32768.0 , a_ : Tuple=0.97 , a_ : List[str]=1.0 , a_ : Dict=True , a_ : str=True , a_ : Union[str, Any]=False , **a_ : int , ): """simple docstring""" super().__init__(feature_size=a_ , sampling_rate=a_ , padding_value=a_ , **a_ ) __snake_case = feature_size __snake_case = sampling_rate __snake_case = padding_value __snake_case = hop_length __snake_case = win_length __snake_case = frame_signal_scale __snake_case = preemphasis_coeff __snake_case = mel_floor __snake_case = normalize_means __snake_case = normalize_vars __snake_case = win_function __snake_case = return_attention_mask __snake_case = win_length * sampling_rate // 1_000 __snake_case = hop_length * sampling_rate // 1_000 __snake_case = optimal_fft_length(self.sample_size ) __snake_case = (self.n_fft // 2) + 1 def A ( self : Optional[Any] , a_ : np.array ): """simple docstring""" if self.win_function == "hamming_window": __snake_case = window_function(window_length=self.sample_size , name=self.win_function , periodic=a_ ) else: __snake_case = window_function(window_length=self.sample_size , name=self.win_function ) __snake_case = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) __snake_case = spectrogram( one_waveform * self.frame_signal_scale , window=a_ , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=a_ , preemphasis=self.preemphasis_coeff , mel_filters=a_ , mel_floor=self.mel_floor , log_mel="log" , ) return msfc_features.T def A ( self : Tuple , a_ : Any , a_ : Tuple , a_ : Dict ): """simple docstring""" if self.normalize_means: __snake_case = x[:input_length].mean(axis=0 ) __snake_case = np.subtract(a_ , a_ ) if self.normalize_vars: __snake_case = x[:input_length].std(axis=0 ) __snake_case = np.divide(a_ , a_ ) if input_length < x.shape[0]: __snake_case = padding_value # make sure array is in float32 __snake_case = x.astype(np.floataa ) return x def A ( self : str , a_ : List[np.ndarray] , a_ : Optional[np.ndarray] = None ): """simple docstring""" __snake_case = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(a_ , a_ , self.padding_value ) for x, n in zip(a_ , a_ )] def __call__( self : List[str] , a_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , a_ : Union[bool, str, PaddingStrategy] = False , a_ : Optional[int] = None , a_ : bool = False , a_ : Optional[int] = None , a_ : Optional[bool] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : Optional[int] = None , **a_ : Optional[int] , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) __snake_case = isinstance(a_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) __snake_case = is_batched_numpy or ( isinstance(a_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __snake_case = [np.asarray(a_ , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a_ , np.ndarray ): __snake_case = np.asarray(a_ , dtype=np.floataa ) elif isinstance(a_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __snake_case = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __snake_case = [raw_speech] # extract fbank features __snake_case = [self._extract_mfsc_features(a_ ) for one_waveform in raw_speech] # convert into correct format for padding __snake_case = BatchFeature({"input_features": features} ) __snake_case = self.pad( a_ , padding=a_ , max_length=a_ , truncation=a_ , pad_to_multiple_of=a_ , return_attention_mask=a_ , **a_ , ) # make sure list is in array format __snake_case = padded_inputs.get("input_features" ) if isinstance(input_features[0] , a_ ): __snake_case = [np.asarray(a_ , dtype=np.floataa ) for feature in input_features] __snake_case = padded_inputs.get("attention_mask" ) if attention_mask is not None: __snake_case = [np.asarray(a_ , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: __snake_case = ( np.array(a_ , dtype=np.intaa ) if self._get_padding_strategies(a_ , max_length=a_ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) __snake_case = self.normalize( padded_inputs["input_features"] , attention_mask=a_ ) if return_tensors is not None: __snake_case = padded_inputs.convert_to_tensors(a_ ) return padded_inputs
69
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : Dict = {'''vocab_file''': '''sentencepiece.model'''} a : Tuple = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } a : str = { '''google/rembert''': 256, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[Any] , a_ : int , a_ : Any=False , a_ : List[Any]=True , a_ : List[Any]=True , a_ : List[Any]="[CLS]" , a_ : List[Any]="[SEP]" , a_ : List[Any]="[UNK]" , a_ : str="[SEP]" , a_ : List[str]="[PAD]" , a_ : Optional[int]="[CLS]" , a_ : List[str]="[MASK]" , **a_ : str , ): """simple docstring""" super().__init__( do_lower_case=a_ , remove_space=a_ , keep_accents=a_ , bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , pad_token=a_ , cls_token=a_ , mask_token=a_ , **a_ , ) __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(a_ ) @property def A ( self : Optional[Any] ): """simple docstring""" return len(self.sp_model ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ): """simple docstring""" __snake_case = self.__dict__.copy() __snake_case = None return state def __setstate__( self : str , a_ : Optional[int] ): """simple docstring""" __snake_case = d __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def A ( self : Tuple , a_ : Optional[int] , a_ : int=False ): """simple docstring""" __snake_case = self.sp_model.EncodeAsPieces(a_ ) return pieces def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" return self.sp_model.PieceToId(a_ ) def A ( self : Optional[Any] , a_ : List[str] ): """simple docstring""" return self.sp_model.IdToPiece(a_ ) def A ( self : Optional[Any] , a_ : int ): """simple docstring""" __snake_case = self.sp_model.decode_pieces(a_ ) return out_string def A ( self : Union[str, Any] , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a_ )) + [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1] def A ( self : Tuple , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A ( self : List[Any] , a_ : str , a_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a_ ): logger.error("Vocabulary path ({}) should be a directory".format(a_ ) ) return __snake_case = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ): copyfile(self.vocab_file , a_ ) return (out_vocab_file,)
69
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = XGLMConfig __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = """gelu""" def __init__( self : int , a_ : List[Any] , a_ : Dict=14 , a_ : int=7 , a_ : List[Any]=True , a_ : Optional[int]=True , a_ : Optional[int]=True , a_ : int=99 , a_ : Optional[int]=32 , a_ : Tuple=2 , a_ : Any=4 , a_ : Union[str, Any]=37 , a_ : Any="gelu" , a_ : str=0.1 , a_ : Any=0.1 , a_ : str=512 , a_ : List[Any]=0.02 , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_labels __snake_case = vocab_size __snake_case = d_model __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = ffn_dim __snake_case = activation_function __snake_case = activation_dropout __snake_case = attention_dropout __snake_case = max_position_embeddings __snake_case = initializer_range __snake_case = None __snake_case = 0 __snake_case = 2 __snake_case = 1 def A ( self : Union[str, Any] ): """simple docstring""" return XGLMConfig.from_pretrained("facebook/xglm-564M" ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = self.get_config() __snake_case = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def A ( self : str ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=a_ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=a_ , ) def A ( self : List[str] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = { "input_ids": input_ids, "head_mask": head_mask, } return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () __SCREAMING_SNAKE_CASE = (TFXGLMForCausalLM,) if is_tf_available() else () __SCREAMING_SNAKE_CASE = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[int] ): """simple docstring""" __snake_case = TFXGLMModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , n_embd=37 ) def A ( self : str ): """simple docstring""" self.config_tester.run_common_tests() @slow def A ( self : Tuple ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = TFXGLMModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) @unittest.skip(reason="Currently, model embeddings are going to undergo a major refactor." ) def A ( self : Dict ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : List[Any] , a_ : Dict=True ): """simple docstring""" __snake_case = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) __snake_case = tf.convert_to_tensor([[2, 268, 9_865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off __snake_case = [2, 268, 9_865, 67, 11, 1_988, 57_252, 9_865, 5, 984, 67, 1_988, 213_838, 1_658, 53, 70_446, 33, 6_657, 278, 1_581] # fmt: on __snake_case = model.generate(a_ , do_sample=a_ , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) __snake_case = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) tf.random.set_seed(0 ) __snake_case = tokenizer("Today is a nice day and" , return_tensors="tf" ) __snake_case = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(":/CPU:0" ): __snake_case = model.generate(a_ , do_sample=a_ , seed=[7, 0] ) __snake_case = tokenizer.decode(output_ids[0] , skip_special_tokens=a_ ) __snake_case = ( "Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due" ) self.assertEqual(a_ , a_ ) @slow def A ( self : List[str] ): """simple docstring""" __snake_case = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) __snake_case = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) __snake_case = "left" # use different length sentences to test batching __snake_case = [ "This is an extremelly long sentence that only exists to test the ability of the model to cope with " "left-padding, such as in batched generation. The output for the sequence below should be the same " "regardless of whether left padding is applied or not. When", "Hello, my dog is a little", ] __snake_case = tokenizer(a_ , return_tensors="tf" , padding=a_ ) __snake_case = inputs["input_ids"] __snake_case = model.generate(input_ids=a_ , attention_mask=inputs["attention_mask"] , max_new_tokens=12 ) __snake_case = tokenizer(sentences[0] , return_tensors="tf" ).input_ids __snake_case = model.generate(input_ids=a_ , max_new_tokens=12 ) __snake_case = tokenizer(sentences[1] , return_tensors="tf" ).input_ids __snake_case = model.generate(input_ids=a_ , max_new_tokens=12 ) __snake_case = tokenizer.batch_decode(a_ , skip_special_tokens=a_ ) __snake_case = tokenizer.decode(output_non_padded[0] , skip_special_tokens=a_ ) __snake_case = tokenizer.decode(output_padded[0] , skip_special_tokens=a_ ) __snake_case = [ "This is an extremelly long sentence that only exists to test the ability of the model to cope with " "left-padding, such as in batched generation. The output for the sequence below should be the same " "regardless of whether left padding is applied or not. When left padding is applied, the sequence will be " "a single", "Hello, my dog is a little bit of a shy one, but he is very friendly", ] self.assertListEqual(a_ , a_ ) self.assertListEqual(a_ , [non_padded_sentence, padded_sentence] )
69
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ): """simple docstring""" try: __snake_case = tempfile.mktemp() with open(a_ , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , a_ ) __snake_case = AlbertTokenizer.from_pretrained(a_ ) finally: os.remove(a_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , a_ ) __snake_case = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A ( self : str ): """simple docstring""" __snake_case = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : List[Any] ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a_ , repo_id="test-tokenizer" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a_ , repo_id="valid_org/test-tokenizer-org" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A ( self : List[str] ): """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizerFast.from_pretrained(a_ ) bert_tokenizer.save_pretrained(a_ ) __snake_case = CustomTokenizerFast.from_pretrained(a_ ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a_ , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A ( self : str ): """simple docstring""" __snake_case = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def A ( self : List[Any] ): """simple docstring""" __snake_case = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : str ): """simple docstring""" __snake_case = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def A ( self : Tuple ): """simple docstring""" __snake_case = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def A ( self : Any ): """simple docstring""" __snake_case = Trie() __snake_case = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a_ , ["AB", "C"] )
69
1
'''simple docstring''' import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) a : int = logging.getLogger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : Tuple ) -> Dict: __snake_case = np.argmax(_UpperCAmelCase , axis=1 ) return np.sum(outputs == labels ) def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Dict: with open(_UpperCAmelCase , encoding="utf_8" ) as f: __snake_case = csv.reader(_UpperCAmelCase ) __snake_case = [] next(_UpperCAmelCase ) # skip the first line for line in tqdm(_UpperCAmelCase ): output.append((" ".join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] ) -> List[Any]: __snake_case = [] for dataset in encoded_datasets: __snake_case = len(_UpperCAmelCase ) __snake_case = np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) __snake_case = np.zeros((n_batch, 2) , dtype=np.intaa ) __snake_case = np.full((n_batch, 2, input_len) , fill_value=-1_00 , dtype=np.intaa ) __snake_case = np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(_UpperCAmelCase ): __snake_case = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __snake_case = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] __snake_case = with_conta __snake_case = with_conta __snake_case = len(_UpperCAmelCase ) - 1 __snake_case = len(_UpperCAmelCase ) - 1 __snake_case = with_conta __snake_case = with_conta __snake_case = mc_label __snake_case = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(_UpperCAmelCase ) for t in all_inputs ) ) return tensor_datasets def __UpperCAmelCase ( ) -> Optional[int]: __snake_case = argparse.ArgumentParser() parser.add_argument("--model_name" , type=_UpperCAmelCase , default="openai-gpt" , help="pretrained model name" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_eval" , action="store_true" , help="Whether to run eval on the dev set." ) parser.add_argument( "--output_dir" , default=_UpperCAmelCase , type=_UpperCAmelCase , required=_UpperCAmelCase , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument("--train_dataset" , type=_UpperCAmelCase , default="" ) parser.add_argument("--eval_dataset" , type=_UpperCAmelCase , default="" ) parser.add_argument("--seed" , type=_UpperCAmelCase , default=42 ) parser.add_argument("--num_train_epochs" , type=_UpperCAmelCase , default=3 ) parser.add_argument("--train_batch_size" , type=_UpperCAmelCase , default=8 ) parser.add_argument("--eval_batch_size" , type=_UpperCAmelCase , default=16 ) parser.add_argument("--adam_epsilon" , default=1E-8 , type=_UpperCAmelCase , help="Epsilon for Adam optimizer." ) parser.add_argument("--max_grad_norm" , type=_UpperCAmelCase , default=1 ) parser.add_argument( "--max_steps" , default=-1 , type=_UpperCAmelCase , help=( "If > 0: set total number of training steps to perform. Override num_train_epochs." ) , ) parser.add_argument( "--gradient_accumulation_steps" , type=_UpperCAmelCase , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--learning_rate" , type=_UpperCAmelCase , default=6.2_5E-5 ) parser.add_argument("--warmup_steps" , default=0 , type=_UpperCAmelCase , help="Linear warmup over warmup_steps." ) parser.add_argument("--lr_schedule" , type=_UpperCAmelCase , default="warmup_linear" ) parser.add_argument("--weight_decay" , type=_UpperCAmelCase , default=0.01 ) parser.add_argument("--lm_coef" , type=_UpperCAmelCase , default=0.9 ) parser.add_argument("--n_valid" , type=_UpperCAmelCase , default=3_74 ) parser.add_argument("--server_ip" , type=_UpperCAmelCase , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=_UpperCAmelCase , default="" , help="Can be used for distant debugging." ) __snake_case = parser.parse_args() print(_UpperCAmelCase ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=_UpperCAmelCase ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) __snake_case = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) __snake_case = torch.cuda.device_count() logger.info("device: {}, n_gpu {}".format(_UpperCAmelCase , _UpperCAmelCase ) ) if not args.do_train and not args.do_eval: raise ValueError("At least one of `do_train` or `do_eval` must be True." ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset __snake_case = ["_start_", "_delimiter_", "_classify_"] __snake_case = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(_UpperCAmelCase ) __snake_case = tokenizer.convert_tokens_to_ids(_UpperCAmelCase ) __snake_case = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(_UpperCAmelCase ) ) model.to(_UpperCAmelCase ) # Load and encode the datasets def tokenize_and_encode(_UpperCAmelCase : Optional[Any] ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(_UpperCAmelCase ) ) elif isinstance(_UpperCAmelCase , _UpperCAmelCase ): return obj return [tokenize_and_encode(_UpperCAmelCase ) for o in obj] logger.info("Encoding dataset..." ) __snake_case = load_rocstories_dataset(args.train_dataset ) __snake_case = load_rocstories_dataset(args.eval_dataset ) __snake_case = (train_dataset, eval_dataset) __snake_case = tokenize_and_encode(_UpperCAmelCase ) # Compute the max input length for the Transformer __snake_case = model.config.n_positions // 2 - 2 __snake_case = max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) __snake_case = min(_UpperCAmelCase , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders __snake_case = pre_process_datasets(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , *_UpperCAmelCase ) __snake_case , __snake_case = tensor_datasets[0], tensor_datasets[1] __snake_case = TensorDataset(*_UpperCAmelCase ) __snake_case = RandomSampler(_UpperCAmelCase ) __snake_case = DataLoader(_UpperCAmelCase , sampler=_UpperCAmelCase , batch_size=args.train_batch_size ) __snake_case = TensorDataset(*_UpperCAmelCase ) __snake_case = SequentialSampler(_UpperCAmelCase ) __snake_case = DataLoader(_UpperCAmelCase , sampler=_UpperCAmelCase , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: __snake_case = args.max_steps __snake_case = args.max_steps // (len(_UpperCAmelCase ) // args.gradient_accumulation_steps) + 1 else: __snake_case = len(_UpperCAmelCase ) // args.gradient_accumulation_steps * args.num_train_epochs __snake_case = list(model.named_parameters() ) __snake_case = ["bias", "LayerNorm.bias", "LayerNorm.weight"] __snake_case = [ { "params": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], "weight_decay": args.weight_decay, }, {"params": [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], "weight_decay": 0.0}, ] __snake_case = AdamW(_UpperCAmelCase , lr=args.learning_rate , eps=args.adam_epsilon ) __snake_case = get_linear_schedule_with_warmup( _UpperCAmelCase , num_warmup_steps=args.warmup_steps , num_training_steps=_UpperCAmelCase ) if args.do_train: __snake_case , __snake_case , __snake_case = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc="Epoch" ): __snake_case = 0 __snake_case = 0 __snake_case = tqdm(_UpperCAmelCase , desc="Training" ) for step, batch in enumerate(_UpperCAmelCase ): __snake_case = tuple(t.to(_UpperCAmelCase ) for t in batch ) __snake_case , __snake_case , __snake_case , __snake_case = batch __snake_case = model(_UpperCAmelCase , mc_token_ids=_UpperCAmelCase , lm_labels=_UpperCAmelCase , mc_labels=_UpperCAmelCase ) __snake_case = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() __snake_case = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 __snake_case = "Training loss: {:.2e} lr: {:.2e}".format(_UpperCAmelCase , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer __snake_case = model.module if hasattr(_UpperCAmelCase , "module" ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` __snake_case = os.path.join(args.output_dir , _UpperCAmelCase ) __snake_case = os.path.join(args.output_dir , _UpperCAmelCase ) torch.save(model_to_save.state_dict() , _UpperCAmelCase ) model_to_save.config.to_json_file(_UpperCAmelCase ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned __snake_case = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) __snake_case = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(_UpperCAmelCase ) if args.do_eval: model.eval() __snake_case , __snake_case = 0, 0 __snake_case , __snake_case = 0, 0 for batch in tqdm(_UpperCAmelCase , desc="Evaluating" ): __snake_case = tuple(t.to(_UpperCAmelCase ) for t in batch ) __snake_case , __snake_case , __snake_case , __snake_case = batch with torch.no_grad(): __snake_case , __snake_case , __snake_case , __snake_case = model( _UpperCAmelCase , mc_token_ids=_UpperCAmelCase , lm_labels=_UpperCAmelCase , mc_labels=_UpperCAmelCase ) __snake_case = mc_logits.detach().cpu().numpy() __snake_case = mc_labels.to("cpu" ).numpy() __snake_case = accuracy(_UpperCAmelCase , _UpperCAmelCase ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 __snake_case = eval_loss / nb_eval_steps __snake_case = eval_accuracy / nb_eval_examples __snake_case = tr_loss / nb_tr_steps if args.do_train else None __snake_case = {"eval_loss": eval_loss, "eval_accuracy": eval_accuracy, "train_loss": train_loss} __snake_case = os.path.join(args.output_dir , "eval_results.txt" ) with open(_UpperCAmelCase , "w" ) as writer: logger.info("***** Eval results *****" ) for key in sorted(result.keys() ): logger.info(" %s = %s" , _UpperCAmelCase , str(result[key] ) ) writer.write("%s = %s\n" % (key, str(result[key] )) ) if __name__ == "__main__": main()
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 __snake_case , __snake_case = 1, 1 for _ in range(number_of_steps - 1 ): __snake_case , __snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import json import os import tempfile from unittest.mock import patch import torch from torch.utils.data import DataLoader, TensorDataset from accelerate import DistributedType, infer_auto_device_map, init_empty_weights from accelerate.accelerator import Accelerator from accelerate.state import GradientState, PartialState from accelerate.test_utils import require_bnb, require_multi_gpu, slow from accelerate.test_utils.testing import AccelerateTestCase, require_cuda from accelerate.utils import patch_environment def __UpperCAmelCase ( ) -> Optional[int]: __snake_case = torch.nn.Linear(2 , 4 ) __snake_case = torch.optim.AdamW(model.parameters() , lr=1.0 ) __snake_case = torch.optim.lr_scheduler.OneCycleLR(_UpperCAmelCase , max_lr=0.01 , steps_per_epoch=2 , epochs=1 ) __snake_case = DataLoader(TensorDataset(torch.tensor([1, 2, 3] ) ) ) __snake_case = DataLoader(TensorDataset(torch.tensor([4, 5, 6] ) ) ) return model, optimizer, scheduler, train_dl, valid_dl def __UpperCAmelCase ( _UpperCAmelCase : List[Any] ) -> List[Any]: return (model.weight.abs().sum() + model.bias.abs().sum()).item() def __UpperCAmelCase ( _UpperCAmelCase : int ) -> Union[str, Any]: __snake_case = torch.nn.Linear(*tuple(model.weight.T.shape ) ).state_dict() model.load_state_dict(_UpperCAmelCase ) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): @require_cuda def A ( self : List[Any] ): """simple docstring""" __snake_case = Accelerator() assert PartialState._shared_state["_cpu"] is False assert PartialState._shared_state["device"].type == "cuda" with self.assertRaises(a_ ): __snake_case = Accelerator(cpu=a_ ) def A ( self : int ): """simple docstring""" __snake_case = Accelerator() __snake_case = GradientState() assert state.num_steps == 1 __snake_case = 4 assert state.num_steps == 4 assert state.sync_gradients is True __snake_case = False assert state.sync_gradients is False GradientState._reset_state() def A ( self : int ): """simple docstring""" __snake_case = Accelerator() __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = create_components() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = accelerator.prepare(a_ , a_ , a_ , a_ , a_ ) self.assertTrue(prepared_model in accelerator._models ) self.assertTrue(prepared_optimizer in accelerator._optimizers ) self.assertTrue(prepared_scheduler in accelerator._schedulers ) self.assertTrue(prepared_train_dl in accelerator._dataloaders ) self.assertTrue(prepared_valid_dl in accelerator._dataloaders ) def A ( self : int ): """simple docstring""" __snake_case = Accelerator() __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = create_components() accelerator.prepare(a_ , a_ , a_ , a_ , a_ ) accelerator.free_memory() self.assertTrue(len(accelerator._models ) == 0 ) self.assertTrue(len(accelerator._optimizers ) == 0 ) self.assertTrue(len(accelerator._schedulers ) == 0 ) self.assertTrue(len(accelerator._dataloaders ) == 0 ) def A ( self : Tuple ): """simple docstring""" PartialState._reset_state() # Mock torch.cuda.set_device to avoid an exception as the device doesn't exist def noop(*a_ : Tuple , **a_ : Optional[int] ): pass with patch("torch.cuda.set_device" , a_ ), patch_environment(ACCELERATE_TORCH_DEVICE="cuda:64" ): __snake_case = Accelerator() self.assertEqual(str(accelerator.state.device ) , "cuda:64" ) def A ( self : List[str] ): """simple docstring""" __snake_case = Accelerator() __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = create_components() accelerator.prepare(a_ , a_ , a_ , a_ , a_ ) __snake_case = get_signature(a_ ) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(a_ ) # make sure random weights don't match load_random_weights(a_ ) self.assertTrue(abs(model_signature - get_signature(a_ ) ) > 1e-3 ) # make sure loaded weights match accelerator.load_state(a_ ) self.assertTrue(abs(model_signature - get_signature(a_ ) ) < 1e-3 ) def A ( self : Dict ): """simple docstring""" __snake_case = Accelerator() __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = create_components() accelerator.prepare(a_ , a_ , a_ , a_ , a_ ) __snake_case = get_signature(a_ ) # saving hook def save_config(a_ : int , a_ : str , a_ : List[str] ): __snake_case = {"class_name": models[0].__class__.__name__} with open(os.path.join(a_ , "data.json" ) , "w" ) as f: json.dump(a_ , a_ ) # loading hook def load_config(a_ : int , a_ : int ): with open(os.path.join(a_ , "data.json" ) , "r" ) as f: __snake_case = json.load(a_ ) __snake_case = config["class_name"] __snake_case = accelerator.register_save_state_pre_hook(a_ ) __snake_case = accelerator.register_load_state_pre_hook(a_ ) with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(a_ ) # make sure random weights don't match with hooks load_random_weights(a_ ) self.assertTrue(abs(model_signature - get_signature(a_ ) ) > 1e-3 ) # random class name to verify correct one is loaded __snake_case = "random" # make sure loaded weights match with hooks accelerator.load_state(a_ ) self.assertTrue(abs(model_signature - get_signature(a_ ) ) < 1e-3 ) # mode.class_name is loaded from config self.assertTrue(model.class_name == model.__class__.__name__ ) # remove hooks save_hook.remove() load_hook.remove() with tempfile.TemporaryDirectory() as tmpdirname: accelerator.save_state(a_ ) # make sure random weights don't match with hooks removed load_random_weights(a_ ) self.assertTrue(abs(model_signature - get_signature(a_ ) ) > 1e-3 ) # random class name to verify correct one is loaded __snake_case = "random" # make sure loaded weights match with hooks removed accelerator.load_state(a_ ) self.assertTrue(abs(model_signature - get_signature(a_ ) ) < 1e-3 ) # mode.class_name is NOT loaded from config self.assertTrue(model.class_name != model.__class__.__name__ ) def A ( self : int ): """simple docstring""" __snake_case = Accelerator() __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = create_components() __snake_case = None # This should work __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = accelerator.prepare( a_ , a_ , a_ , a_ , a_ , a_ ) self.assertTrue(dummy_obj is None ) def A ( self : List[str] ): """simple docstring""" __snake_case = Accelerator() __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = create_components() __snake_case = [1, 2, 3] # This should work __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case = accelerator.prepare( a_ , a_ , a_ , a_ , a_ , a_ ) self.assertEqual( getattr(a_ , "_is_accelerate_prepared" , a_ ) , a_ , "Dummy object should have `_is_accelerate_prepared` set to `True`" , ) self.assertEqual( getattr(a_ , "_is_accelerate_prepared" , a_ ) , a_ , "Model is missing `_is_accelerator_prepared` or is set to `False`" , ) self.assertEqual( getattr(a_ , "_is_accelerate_prepared" , a_ ) , a_ , "Optimizer is missing `_is_accelerator_prepared` or is set to `False`" , ) self.assertEqual( getattr(a_ , "_is_accelerate_prepared" , a_ ) , a_ , "Scheduler is missing `_is_accelerator_prepared` or is set to `False`" , ) self.assertEqual( getattr(a_ , "_is_accelerate_prepared" , a_ ) , a_ , "Train Dataloader is missing `_is_accelerator_prepared` or is set to `False`" , ) self.assertEqual( getattr(a_ , "_is_accelerate_prepared" , a_ ) , a_ , "Valid Dataloader is missing `_is_accelerator_prepared` or is set to `False`" , ) @slow @require_bnb def A ( self : Union[str, Any] ): """simple docstring""" from transformers import AutoModelForCausalLM __snake_case = AutoModelForCausalLM.from_pretrained( "EleutherAI/gpt-neo-125m" , load_in_abit=a_ , device_map={"": 0} , ) __snake_case = Accelerator() # This should work __snake_case = accelerator.prepare(a_ ) @slow @require_bnb def A ( self : List[Any] ): """simple docstring""" from transformers import AutoModelForCausalLM __snake_case = Accelerator() with init_empty_weights(): __snake_case = AutoModelForCausalLM.from_pretrained( "EleutherAI/gpt-neo-125m" , ) model.tie_weights() __snake_case = infer_auto_device_map(a_ ) __snake_case = "cpu" __snake_case = AutoModelForCausalLM.from_pretrained( "EleutherAI/gpt-neo-125m" , device_map=a_ , load_in_abit=a_ , llm_inta_enable_fpaa_cpu_offload=a_ ) # This should not work and get value error with self.assertRaises(a_ ): __snake_case = accelerator.prepare(a_ ) @slow @require_bnb @require_multi_gpu def A ( self : Optional[Any] ): """simple docstring""" from transformers import AutoModelForCausalLM __snake_case = {"distributed_type": DistributedType.MULTI_GPU} with init_empty_weights(): __snake_case = AutoModelForCausalLM.from_pretrained( "EleutherAI/gpt-neo-125m" , ) model.tie_weights() __snake_case = infer_auto_device_map(a_ ) __snake_case = 1 __snake_case = AutoModelForCausalLM.from_pretrained( "EleutherAI/gpt-neo-125m" , load_in_abit=a_ , device_map=a_ , ) __snake_case = Accelerator() # This should not work and get value error with self.assertRaises(a_ ): __snake_case = accelerator.prepare(a_ ) PartialState._reset_state() @slow @require_bnb @require_multi_gpu def A ( self : Tuple ): """simple docstring""" from transformers import AutoModelForCausalLM with init_empty_weights(): __snake_case = AutoModelForCausalLM.from_pretrained( "EleutherAI/gpt-neo-125m" , ) __snake_case = infer_auto_device_map(a_ ) __snake_case = 1 __snake_case = AutoModelForCausalLM.from_pretrained( "EleutherAI/gpt-neo-125m" , load_in_abit=a_ , device_map=a_ , ) __snake_case = Accelerator() # This should work __snake_case = accelerator.prepare(a_ ) @require_cuda def A ( self : Optional[int] ): """simple docstring""" __snake_case = torch.nn.Linear(10 , 10 ) __snake_case = torch.optim.SGD(model.parameters() , lr=0.01 ) __snake_case = Accelerator(cpu=a_ ) __snake_case = accelerator.prepare(a_ )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str ) -> str: return " ".join( "".join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
69
1
'''simple docstring''' import argparse import glob import logging import os from argparse import Namespace from importlib import import_module import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch.nn import CrossEntropyLoss from torch.utils.data import DataLoader, TensorDataset from utils_ner import TokenClassificationTask a : Optional[Any] = logging.getLogger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """token-classification""" def __init__( self : str , a_ : Tuple ): """simple docstring""" if type(a_ ) == dict: __snake_case = Namespace(**a_ ) __snake_case = import_module("tasks" ) try: __snake_case = getattr(a_ , hparams.task_type ) __snake_case = token_classification_task_clazz() except AttributeError: raise ValueError( f'''Task {hparams.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. ''' f'''Available tasks classes are: {TokenClassificationTask.__subclasses__()}''' ) __snake_case = self.token_classification_task.get_labels(hparams.labels ) __snake_case = CrossEntropyLoss().ignore_index super().__init__(a_ , len(self.labels ) , self.mode ) def A ( self : int , **a_ : List[str] ): """simple docstring""" return self.model(**a_ ) def A ( self : str , a_ : List[str] , a_ : Dict ): """simple docstring""" __snake_case = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": __snake_case = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids __snake_case = self(**a_ ) __snake_case = outputs[0] # tensorboard_logs = {"loss": loss, "rate": self.lr_scheduler.get_last_lr()[-1]} return {"loss": loss} def A ( self : List[Any] ): """simple docstring""" __snake_case = self.hparams for mode in ["train", "dev", "test"]: __snake_case = self._feature_file(a_ ) if os.path.exists(a_ ) and not args.overwrite_cache: logger.info("Loading features from cached file %s" , a_ ) __snake_case = torch.load(a_ ) else: logger.info("Creating features from dataset file at %s" , args.data_dir ) __snake_case = self.token_classification_task.read_examples_from_file(args.data_dir , a_ ) __snake_case = self.token_classification_task.convert_examples_to_features( a_ , self.labels , args.max_seq_length , self.tokenizer , cls_token_at_end=bool(self.config.model_type in ["xlnet"] ) , cls_token=self.tokenizer.cls_token , cls_token_segment_id=2 if self.config.model_type in ["xlnet"] else 0 , sep_token=self.tokenizer.sep_token , sep_token_extra=a_ , pad_on_left=bool(self.config.model_type in ["xlnet"] ) , pad_token=self.tokenizer.pad_token_id , pad_token_segment_id=self.tokenizer.pad_token_type_id , pad_token_label_id=self.pad_token_label_id , ) logger.info("Saving features into cached file %s" , a_ ) torch.save(a_ , a_ ) def A ( self : List[Any] , a_ : int , a_ : int , a_ : bool = False ): """simple docstring""" __snake_case = self._feature_file(a_ ) logger.info("Loading features from cached file %s" , a_ ) __snake_case = torch.load(a_ ) __snake_case = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) __snake_case = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) if features[0].token_type_ids is not None: __snake_case = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) else: __snake_case = torch.tensor([0 for f in features] , dtype=torch.long ) # HACK(we will not use this anymore soon) __snake_case = torch.tensor([f.label_ids for f in features] , dtype=torch.long ) return DataLoader( TensorDataset(a_ , a_ , a_ , a_ ) , batch_size=a_ ) def A ( self : Tuple , a_ : Tuple , a_ : Optional[Any] ): """simple docstring""" """Compute validation""" "" __snake_case = {"input_ids": batch[0], "attention_mask": batch[1], "labels": batch[3]} if self.config.model_type != "distilbert": __snake_case = ( batch[2] if self.config.model_type in ["bert", "xlnet"] else None ) # XLM and RoBERTa don"t use token_type_ids __snake_case = self(**a_ ) __snake_case , __snake_case = outputs[:2] __snake_case = logits.detach().cpu().numpy() __snake_case = inputs["labels"].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def A ( self : Dict , a_ : Tuple ): """simple docstring""" __snake_case = torch.stack([x["val_loss"] for x in outputs] ).mean() __snake_case = np.concatenate([x["pred"] for x in outputs] , axis=0 ) __snake_case = np.argmax(a_ , axis=2 ) __snake_case = np.concatenate([x["target"] for x in outputs] , axis=0 ) __snake_case = dict(enumerate(self.labels ) ) __snake_case = [[] for _ in range(out_label_ids.shape[0] )] __snake_case = [[] for _ in range(out_label_ids.shape[0] )] for i in range(out_label_ids.shape[0] ): for j in range(out_label_ids.shape[1] ): if out_label_ids[i, j] != self.pad_token_label_id: out_label_list[i].append(label_map[out_label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) __snake_case = { "val_loss": val_loss_mean, "accuracy_score": accuracy_score(a_ , a_ ), "precision": precision_score(a_ , a_ ), "recall": recall_score(a_ , a_ ), "f1": fa_score(a_ , a_ ), } __snake_case = dict(results.items() ) __snake_case = results return ret, preds_list, out_label_list def A ( self : List[str] , a_ : List[Any] ): """simple docstring""" __snake_case , __snake_case , __snake_case = self._eval_end(a_ ) __snake_case = ret["log"] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def A ( self : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case , __snake_case , __snake_case = self._eval_end(a_ ) # Converting to the dict required by pl # https://github.com/PyTorchLightning/pytorch-lightning/blob/master/\ # pytorch_lightning/trainer/logging.py#L139 __snake_case = ret["log"] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def A ( a_ : int , a_ : List[str] ): """simple docstring""" BaseTransformer.add_model_specific_args(a_ , a_ ) parser.add_argument( "--task_type" , default="NER" , type=a_ , help="Task type to fine tune in training (e.g. NER, POS, etc)" ) parser.add_argument( "--max_seq_length" , default=128 , type=a_ , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--labels" , default="" , type=a_ , help="Path to a file containing all labels. If not specified, CoNLL-2003 labels are used." , ) parser.add_argument( "--gpus" , default=0 , type=a_ , help="The number of GPUs allocated for this, it is by default 0 meaning none" , ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) return parser if __name__ == "__main__": a : Union[str, Any] = argparse.ArgumentParser() add_generic_args(parser, os.getcwd()) a : List[str] = NERTransformer.add_model_specific_args(parser, os.getcwd()) a : List[Any] = parser.parse_args() a : str = NERTransformer(args) a : List[Any] = generic_train(model, args) if args.do_predict: # See https://github.com/huggingface/transformers/issues/3159 # pl use this default format to create a checkpoint: # https://github.com/PyTorchLightning/pytorch-lightning/blob/master\ # /pytorch_lightning/callbacks/model_checkpoint.py#L322 a : Tuple = sorted(glob.glob(os.path.join(args.output_dir, '''checkpoint-epoch=*.ckpt'''), recursive=True)) a : Optional[int] = model.load_from_checkpoint(checkpoints[-1]) trainer.test(model)
69
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : Any , a_ : Union[str, Any]=13 , a_ : Any=7 , a_ : Any=True , a_ : Dict=True , a_ : Union[str, Any]=False , a_ : Tuple=True , a_ : str=99 , a_ : Tuple=64 , a_ : Tuple=5 , a_ : Union[str, Any]=4 , a_ : Dict=64 , a_ : Union[str, Any]="gelu" , a_ : Dict=0.1 , a_ : List[str]=0.1 , a_ : Dict=512 , a_ : Tuple=16 , a_ : str=2 , a_ : Any=0.02 , a_ : List[Any]=3 , a_ : Tuple=4 , a_ : Optional[int]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : int ): """simple docstring""" return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def A ( self : str ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def A ( self : Tuple , a_ : int , a_ : str , a_ : Optional[int] , a_ : List[Any] , a_ : str , a_ : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Any , a_ : int , a_ : Tuple , a_ : str , a_ : int , a_ : str , a_ : List[Any] ): """simple docstring""" __snake_case = MPNetForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Any , a_ : Any , a_ : int , a_ : Union[str, Any] , a_ : Dict , a_ : Optional[Any] , a_ : Any ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Optional[Any] , a_ : Any , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = MPNetForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict , a_ : List[str] , a_ : str , a_ : Union[str, Any] , a_ : str , a_ : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def A ( self : List[Any] ): """simple docstring""" __snake_case = MPNetModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel.from_pretrained("microsoft/mpnet-base" ) __snake_case = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a_ , atol=1e-4 ) )
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 50 ) -> int: __snake_case = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(F'''{solution() = }''')
69
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Optional[int]: return 1 / (1 + np.exp(-z )) def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict ) -> List[str]: return (-y * np.log(_UpperCAmelCase ) - (1 - y) * np.log(1 - h )).mean() def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any] ) -> Optional[Any]: __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) return np.sum(y * scores - np.log(1 + np.exp(_UpperCAmelCase ) ) ) def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str]=7_00_00 ) -> Union[str, Any]: __snake_case = np.zeros(x.shape[1] ) for iterations in range(_UpperCAmelCase ): __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = np.dot(x.T , h - y ) / y.size __snake_case = theta - alpha * gradient # updating the weights __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = cost_function(_UpperCAmelCase , _UpperCAmelCase ) if iterations % 1_00 == 0: print(F'''loss: {j} \t''' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": a : int = datasets.load_iris() a : int = iris.data[:, :2] a : Optional[Any] = (iris.target != 0) * 1 a : Tuple = 0.1 a : List[str] = logistic_reg(alpha, x, y, max_iterations=70_000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> Union[str, Any]: return sigmoid_function( np.dot(_UpperCAmelCase , _UpperCAmelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((a) , (a)) : Any = (x[:, 0].min(), x[:, 0].max()) ((a) , (a)) : Any = (x[:, 1].min(), x[:, 1].max()) ((a) , (a)) : Any = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) a : Optional[Any] = np.c_[xxa.ravel(), xxa.ravel()] a : List[Any] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
69
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer a : Optional[int] = logging.get_logger(__name__) a : List[Any] = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} # See all MVP models at https://huggingface.co/models?filter=mvp a : str = { '''vocab_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json''', }, '''added_tokens.json''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json''', }, '''merges_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''RUCAIBox/mvp''': '''https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json''', }, } a : Optional[int] = { '''RUCAIBox/mvp''': 1_024, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE = ["""input_ids""", """attention_mask"""] __SCREAMING_SNAKE_CASE = MvpTokenizer def __init__( self : Optional[int] , a_ : Tuple=None , a_ : Optional[int]=None , a_ : List[str]=None , a_ : int="replace" , a_ : Tuple="<s>" , a_ : str="</s>" , a_ : Tuple="</s>" , a_ : Dict="<s>" , a_ : str="<unk>" , a_ : Optional[int]="<pad>" , a_ : Optional[int]="<mask>" , a_ : Optional[Any]=False , a_ : Union[str, Any]=True , **a_ : Optional[int] , ): """simple docstring""" super().__init__( a_ , a_ , tokenizer_file=a_ , errors=a_ , bos_token=a_ , eos_token=a_ , sep_token=a_ , cls_token=a_ , unk_token=a_ , pad_token=a_ , mask_token=a_ , add_prefix_space=a_ , trim_offsets=a_ , **a_ , ) __snake_case = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , a_ ) != add_prefix_space: __snake_case = getattr(a_ , pre_tok_state.pop("type" ) ) __snake_case = add_prefix_space __snake_case = pre_tok_class(**a_ ) __snake_case = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` __snake_case = "post_processor" __snake_case = getattr(self.backend_tokenizer , a_ , a_ ) if tokenizer_component_instance: __snake_case = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __snake_case = tuple(state["sep"] ) if "cls" in state: __snake_case = tuple(state["cls"] ) __snake_case = False if state.get("add_prefix_space" , a_ ) != add_prefix_space: __snake_case = add_prefix_space __snake_case = True if state.get("trim_offsets" , a_ ) != trim_offsets: __snake_case = trim_offsets __snake_case = True if changes_to_apply: __snake_case = getattr(a_ , state.pop("type" ) ) __snake_case = component_class(**a_ ) setattr(self.backend_tokenizer , a_ , a_ ) @property def A ( self : Optional[Any] ): """simple docstring""" if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A ( self : str , a_ : Any ): """simple docstring""" __snake_case = AddedToken(a_ , lstrip=a_ , rstrip=a_ ) if isinstance(a_ , a_ ) else value __snake_case = value def A ( self : int , *a_ : Dict , **a_ : int ): """simple docstring""" __snake_case = kwargs.get("is_split_into_words" , a_ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*a_ , **a_ ) def A ( self : Optional[Any] , *a_ : Dict , **a_ : str ): """simple docstring""" __snake_case = kwargs.get("is_split_into_words" , a_ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*a_ , **a_ ) def A ( self : int , a_ : str , a_ : Optional[str] = None ): """simple docstring""" __snake_case = self._tokenizer.model.save(a_ , name=a_ ) return tuple(a_ ) def A ( self : List[str] , a_ : Optional[int] , a_ : Optional[int]=None ): """simple docstring""" __snake_case = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A ( self : str , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder a : int = '''__DUMMY_TRANSFORMERS_USER__''' a : str = '''Dummy User''' a : Dict = '''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''' a : Optional[Any] = '''https://hub-ci.huggingface.co''' a : Any = CI_HUB_ENDPOINT + '''/datasets/{repo_id}/resolve/{revision}/{path}''' a : Any = CI_HUB_ENDPOINT + '''/{repo_id}/resolve/{revision}/{filename}''' a : Union[str, Any] = Path('''~/.huggingface/hub_ci_token''').expanduser() @pytest.fixture def __UpperCAmelCase ( _UpperCAmelCase : Tuple ) -> Dict: monkeypatch.setattr( "huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE" , _UpperCAmelCase ) @pytest.fixture def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] ) -> Union[str, Any]: monkeypatch.setattr("datasets.config.HF_ENDPOINT" , _UpperCAmelCase ) monkeypatch.setattr("datasets.config.HUB_DATASETS_URL" , _UpperCAmelCase ) @pytest.fixture def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Dict: monkeypatch.setattr("huggingface_hub.hf_api.HfFolder.path_token" , _UpperCAmelCase ) @pytest.fixture def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : Dict ) -> Dict: HfFolder.save_token(_UpperCAmelCase ) yield HfFolder.delete_token() @pytest.fixture(scope="session" ) def __UpperCAmelCase ( ) -> Union[str, Any]: return HfApi(endpoint=_UpperCAmelCase ) @pytest.fixture(scope="session" ) def __UpperCAmelCase ( _UpperCAmelCase : HfApi ) -> List[Any]: __snake_case = HfFolder.get_token() HfFolder.save_token(_UpperCAmelCase ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(_UpperCAmelCase ) @pytest.fixture def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] ) -> int: def _cleanup_repo(_UpperCAmelCase : Optional[int] ): hf_api.delete_repo(_UpperCAmelCase , token=_UpperCAmelCase , repo_type="dataset" ) return _cleanup_repo @pytest.fixture def __UpperCAmelCase ( _UpperCAmelCase : Tuple ) -> Optional[int]: @contextmanager def _temporary_repo(_UpperCAmelCase : str ): try: yield repo_id finally: cleanup_repo(_UpperCAmelCase ) return _temporary_repo @pytest.fixture(scope="session" ) def __UpperCAmelCase ( _UpperCAmelCase : HfApi , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[str] ) -> List[str]: __snake_case = F'''repo_txt_data-{int(time.time() * 1_0E3 )}''' __snake_case = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_UpperCAmelCase , token=_UpperCAmelCase , repo_type="dataset" , private=_UpperCAmelCase ) hf_api.upload_file( token=_UpperCAmelCase , path_or_fileobj=str(_UpperCAmelCase ) , path_in_repo="data/text_data.txt" , repo_id=_UpperCAmelCase , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(_UpperCAmelCase , token=_UpperCAmelCase , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int ) -> Optional[int]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="session" ) def __UpperCAmelCase ( _UpperCAmelCase : HfApi , _UpperCAmelCase : int , _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = F'''repo_zipped_txt_data-{int(time.time() * 1_0E3 )}''' __snake_case = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_UpperCAmelCase , token=_UpperCAmelCase , repo_type="dataset" , private=_UpperCAmelCase ) hf_api.upload_file( token=_UpperCAmelCase , path_or_fileobj=str(_UpperCAmelCase ) , path_in_repo="data.zip" , repo_id=_UpperCAmelCase , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(_UpperCAmelCase , token=_UpperCAmelCase , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] ) -> Tuple: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="session" ) def __UpperCAmelCase ( _UpperCAmelCase : HfApi , _UpperCAmelCase : Any , _UpperCAmelCase : Any ) -> List[Any]: __snake_case = F'''repo_zipped_img_data-{int(time.time() * 1_0E3 )}''' __snake_case = F'''{CI_HUB_USER}/{repo_name}''' hf_api.create_repo(_UpperCAmelCase , token=_UpperCAmelCase , repo_type="dataset" , private=_UpperCAmelCase ) hf_api.upload_file( token=_UpperCAmelCase , path_or_fileobj=str(_UpperCAmelCase ) , path_in_repo="data.zip" , repo_id=_UpperCAmelCase , repo_type="dataset" , ) yield repo_id try: hf_api.delete_repo(_UpperCAmelCase , token=_UpperCAmelCase , repo_type="dataset" ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __UpperCAmelCase ( _UpperCAmelCase : Any , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str] ) -> Dict: return hf_private_dataset_repo_zipped_img_data_
69
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
1
'''simple docstring''' import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class SCREAMING_SNAKE_CASE__ : @staticmethod def A ( *a_ : Optional[Any] , **a_ : int ): """simple docstring""" pass @is_pipeline_test @require_vision class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @require_torch def A ( self : int ): """simple docstring""" __snake_case = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , ) __snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) __snake_case = image_classifier(a_ , candidate_labels=["a", "b", "c"] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(a_ ) , [ [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "c"}, {"score": 0.333, "label": "b"}], ] , ) __snake_case = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(a_ ) , [ [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], ] , ) @require_tf def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , framework="tf" ) __snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) __snake_case = image_classifier(a_ , candidate_labels=["a", "b", "c"] ) self.assertEqual( nested_simplify(a_ ) , [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}] , ) __snake_case = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(a_ ) , [ [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], [ {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, {"score": 0.333, "label": ANY(a_ )}, ], ] , ) @slow @require_torch def A ( self : Tuple ): """simple docstring""" __snake_case = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , ) # This is an image of 2 cats with remotes and no planes __snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) __snake_case = image_classifier(a_ , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(a_ ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) __snake_case = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(a_ ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , ) @slow @require_tf def A ( self : int ): """simple docstring""" __snake_case = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , framework="tf" ) # This is an image of 2 cats with remotes and no planes __snake_case = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) __snake_case = image_classifier(a_ , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(a_ ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) __snake_case = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(a_ ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , )
69
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Tuple = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """autoformer""" __SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : List[Any] , a_ : Optional[int] = None , a_ : Optional[int] = None , a_ : str = "student_t" , a_ : str = "nll" , a_ : int = 1 , a_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , a_ : bool = True , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : Optional[List[int]] = None , a_ : Optional[List[int]] = None , a_ : int = 64 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 32 , a_ : int = 32 , a_ : str = "gelu" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : int = 100 , a_ : float = 0.02 , a_ : bool = True , a_ : Union[str, Any]=True , a_ : int = 10 , a_ : int = 25 , a_ : int = 3 , **a_ : Tuple , ): """simple docstring""" __snake_case = prediction_length __snake_case = context_length if context_length is not None else prediction_length __snake_case = distribution_output __snake_case = loss __snake_case = input_size __snake_case = num_time_features __snake_case = lags_sequence __snake_case = scaling __snake_case = num_dynamic_real_features __snake_case = num_static_real_features __snake_case = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case = cardinality else: __snake_case = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case = embedding_dimension else: __snake_case = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case = num_parallel_samples # Transformer architecture configuration __snake_case = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case = d_model __snake_case = encoder_attention_heads __snake_case = decoder_attention_heads __snake_case = encoder_ffn_dim __snake_case = decoder_ffn_dim __snake_case = encoder_layers __snake_case = decoder_layers __snake_case = dropout __snake_case = attention_dropout __snake_case = activation_dropout __snake_case = encoder_layerdrop __snake_case = decoder_layerdrop __snake_case = activation_function __snake_case = init_std __snake_case = use_cache # Autoformer __snake_case = label_length __snake_case = moving_average __snake_case = autocorrelation_factor super().__init__(is_encoder_decoder=a_ , **a_ ) @property def A ( self : Optional[int] ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
69
1
'''simple docstring''' import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ConsistencyModelPipeline __SCREAMING_SNAKE_CASE = UNCONDITIONAL_IMAGE_GENERATION_PARAMS __SCREAMING_SNAKE_CASE = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt __SCREAMING_SNAKE_CASE = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def A ( self : int ): """simple docstring""" __snake_case = UNetaDModel.from_pretrained( "diffusers/consistency-models-test" , subfolder="test_unet" , ) return unet @property def A ( self : List[str] ): """simple docstring""" __snake_case = UNetaDModel.from_pretrained( "diffusers/consistency-models-test" , subfolder="test_unet_class_cond" , ) return unet def A ( self : Union[str, Any] , a_ : Dict=False ): """simple docstring""" if class_cond: __snake_case = self.dummy_cond_unet else: __snake_case = self.dummy_uncond_unet # Default to CM multistep sampler __snake_case = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __snake_case = { "unet": unet, "scheduler": scheduler, } return components def A ( self : List[Any] , a_ : Tuple , a_ : Dict=0 ): """simple docstring""" if str(a_ ).startswith("mps" ): __snake_case = torch.manual_seed(a_ ) else: __snake_case = torch.Generator(device=a_ ).manual_seed(a_ ) __snake_case = { "batch_size": 1, "num_inference_steps": None, "timesteps": [22, 0], "generator": generator, "output_type": "np", } return inputs def A ( self : List[str] ): """simple docstring""" __snake_case = "cpu" # ensure determinism for the device-dependent torch.Generator __snake_case = self.get_dummy_components() __snake_case = ConsistencyModelPipeline(**a_ ) __snake_case = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_dummy_inputs(a_ ) __snake_case = pipe(**a_ ).images assert image.shape == (1, 32, 32, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "cpu" # ensure determinism for the device-dependent torch.Generator __snake_case = self.get_dummy_components(class_cond=a_ ) __snake_case = ConsistencyModelPipeline(**a_ ) __snake_case = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_dummy_inputs(a_ ) __snake_case = 0 __snake_case = pipe(**a_ ).images assert image.shape == (1, 32, 32, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def A ( self : Optional[Any] ): """simple docstring""" __snake_case = "cpu" # ensure determinism for the device-dependent torch.Generator __snake_case = self.get_dummy_components() __snake_case = ConsistencyModelPipeline(**a_ ) __snake_case = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_dummy_inputs(a_ ) __snake_case = 1 __snake_case = None __snake_case = pipe(**a_ ).images assert image.shape == (1, 32, 32, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def A ( self : Dict ): """simple docstring""" __snake_case = "cpu" # ensure determinism for the device-dependent torch.Generator __snake_case = self.get_dummy_components(class_cond=a_ ) __snake_case = ConsistencyModelPipeline(**a_ ) __snake_case = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_dummy_inputs(a_ ) __snake_case = 1 __snake_case = None __snake_case = 0 __snake_case = pipe(**a_ ).images assert image.shape == (1, 32, 32, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Dict ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : int , a_ : Dict=0 , a_ : int=False , a_ : Dict="cpu" , a_ : Any=torch.floataa , a_ : Optional[int]=(1, 3, 64, 64) ): """simple docstring""" __snake_case = torch.manual_seed(a_ ) __snake_case = { "num_inference_steps": None, "timesteps": [22, 0], "class_labels": 0, "generator": generator, "output_type": "np", } if get_fixed_latents: __snake_case = self.get_fixed_latents(seed=a_ , device=a_ , dtype=a_ , shape=a_ ) __snake_case = latents return inputs def A ( self : Any , a_ : Optional[int]=0 , a_ : Tuple="cpu" , a_ : List[Any]=torch.floataa , a_ : Optional[Any]=(1, 3, 64, 64) ): """simple docstring""" if type(a_ ) == str: __snake_case = torch.device(a_ ) __snake_case = torch.Generator(device=a_ ).manual_seed(a_ ) __snake_case = randn_tensor(a_ , generator=a_ , device=a_ , dtype=a_ ) return latents def A ( self : Optional[Any] ): """simple docstring""" __snake_case = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) __snake_case = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __snake_case = ConsistencyModelPipeline(unet=a_ , scheduler=a_ ) pipe.to(torch_device=a_ ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_inputs() __snake_case = pipe(**a_ ).images assert image.shape == (1, 64, 64, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.0888, 0.0881, 0.0666, 0.0479, 0.0292, 0.0195, 0.0201, 0.0163, 0.0254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) __snake_case = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __snake_case = ConsistencyModelPipeline(unet=a_ , scheduler=a_ ) pipe.to(torch_device=a_ ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_inputs() __snake_case = 1 __snake_case = None __snake_case = pipe(**a_ ).images assert image.shape == (1, 64, 64, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.0340, 0.0152, 0.0063, 0.0267, 0.0221, 0.0107, 0.0416, 0.0186, 0.0217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 @require_torch_a def A ( self : str ): """simple docstring""" __snake_case = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) __snake_case = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __snake_case = ConsistencyModelPipeline(unet=a_ , scheduler=a_ ) pipe.to(torch_device=a_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_inputs(get_fixed_latents=a_ , device=a_ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a_ , enable_math=a_ , enable_mem_efficient=a_ ): __snake_case = pipe(**a_ ).images assert image.shape == (1, 64, 64, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.1875, 0.1428, 0.1289, 0.2151, 0.2092, 0.1477, 0.1877, 0.1641, 0.1353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 @require_torch_a def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = UNetaDModel.from_pretrained("diffusers/consistency_models" , subfolder="diffusers_cd_imagenet64_l2" ) __snake_case = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __snake_case = ConsistencyModelPipeline(unet=a_ , scheduler=a_ ) pipe.to(torch_device=a_ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = self.get_inputs(get_fixed_latents=a_ , device=a_ ) __snake_case = 1 __snake_case = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=a_ , enable_math=a_ , enable_mem_efficient=a_ ): __snake_case = pipe(**a_ ).images assert image.shape == (1, 64, 64, 3) __snake_case = image[0, -3:, -3:, -1] __snake_case = np.array([0.1663, 0.1948, 0.2275, 0.1680, 0.1204, 0.1245, 0.1858, 0.1338, 0.2095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
69
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
1
'''simple docstring''' import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase , _UpperCamelCase ): def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = load_tool("text-to-speech" ) self.tool.setup() def A ( self : List[str] ): """simple docstring""" torch.manual_seed(0 ) __snake_case = self.tool("hey" ) __snake_case = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485] ) , ) ) def A ( self : int ): """simple docstring""" torch.manual_seed(0 ) __snake_case = self.tool("hey" ) __snake_case = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485] ) , ) )
69
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 a : Tuple = get_tests_dir('''fixtures''') a : Dict = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : int = get_tests_dir('''fixtures/dummy-config.json''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Tuple ): """simple docstring""" __snake_case = 0 def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __snake_case = AutoFeatureExtractor.from_pretrained(a_ ).to_dict() config_dict.pop("feature_extractor_type" ) __snake_case = WavaVecaFeatureExtractor(**a_ ) # save in new folder model_config.save_pretrained(a_ ) config.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) # make sure private variable is not incorrectly saved __snake_case = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with self.assertRaisesRegex( a_ , "bert-base is not a local folder and is not a valid model identifier" ): __snake_case = AutoFeatureExtractor.from_pretrained("bert-base" ) def A ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( a_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __snake_case = AutoFeatureExtractor.from_pretrained(a_ , revision="aaaaaa" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( a_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): __snake_case = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ , trust_remote_code=a_ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def A ( self : int ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoFeatureExtractor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def A ( self : Dict ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # If remote code is not set, the default is to use local __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(a_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
69
1
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('''4.31.0''') a : Any = logging.getLogger(__name__) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) __SCREAMING_SNAKE_CASE = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = field(default=_UpperCamelCase , metadata={"""help""": """The input training data file (a text file)."""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) __SCREAMING_SNAKE_CASE = field( default=_UpperCamelCase , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def A ( self : Any ): """simple docstring""" if self.train_file is not None: __snake_case = self.train_file.split("." )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: __snake_case = self.validation_file.split("." )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = None def __call__( self : List[Any] , a_ : str ): """simple docstring""" __snake_case = "label" if "label" in features[0].keys() else "labels" __snake_case = [feature.pop(a_ ) for feature in features] __snake_case = len(a_ ) __snake_case = len(features[0]["input_ids"] ) __snake_case = [ [{k: v[i] for k, v in feature.items()} for i in range(a_ )] for feature in features ] __snake_case = list(chain(*a_ ) ) __snake_case = self.tokenizer.pad( a_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" , ) # Un-flatten __snake_case = {k: v.view(a_ , a_ , -1 ) for k, v in batch.items()} # Add back labels __snake_case = torch.tensor(a_ , dtype=torch.intaa ) return batch def __UpperCAmelCase ( ) -> Tuple: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. __snake_case , __snake_case , __snake_case = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: __snake_case , __snake_case , __snake_case = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("run_swag" , _UpperCAmelCase , _UpperCAmelCase ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() __snake_case = training_args.get_process_log_level() logger.setLevel(_UpperCAmelCase ) datasets.utils.logging.set_verbosity(_UpperCAmelCase ) transformers.utils.logging.set_verbosity(_UpperCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. __snake_case = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: __snake_case = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: __snake_case = {} if data_args.train_file is not None: __snake_case = data_args.train_file if data_args.validation_file is not None: __snake_case = data_args.validation_file __snake_case = data_args.train_file.split("." )[-1] __snake_case = load_dataset( _UpperCAmelCase , data_files=_UpperCAmelCase , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. __snake_case = load_dataset( "swag" , "regular" , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __snake_case = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) __snake_case = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) __snake_case = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=_UpperCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. __snake_case = [F'''ending{i}''' for i in range(4 )] __snake_case = "sent1" __snake_case = "sent2" if data_args.max_seq_length is None: __snake_case = tokenizer.model_max_length if max_seq_length > 10_24: logger.warning( "The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value" " of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can" " override this default with `--block_size xxx`." ) __snake_case = 10_24 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) __snake_case = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(_UpperCAmelCase : Union[str, Any] ): __snake_case = [[context] * 4 for context in examples[context_name]] __snake_case = examples[question_header_name] __snake_case = [ [F'''{header} {examples[end][i]}''' for end in ending_names] for i, header in enumerate(_UpperCAmelCase ) ] # Flatten out __snake_case = list(chain(*_UpperCAmelCase ) ) __snake_case = list(chain(*_UpperCAmelCase ) ) # Tokenize __snake_case = tokenizer( _UpperCAmelCase , _UpperCAmelCase , truncation=_UpperCAmelCase , max_length=_UpperCAmelCase , padding="max_length" if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(_UpperCAmelCase ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError("--do_train requires a train dataset" ) __snake_case = raw_datasets["train"] if data_args.max_train_samples is not None: __snake_case = min(len(_UpperCAmelCase ) , data_args.max_train_samples ) __snake_case = train_dataset.select(range(_UpperCAmelCase ) ) with training_args.main_process_first(desc="train dataset map pre-processing" ): __snake_case = train_dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError("--do_eval requires a validation dataset" ) __snake_case = raw_datasets["validation"] if data_args.max_eval_samples is not None: __snake_case = min(len(_UpperCAmelCase ) , data_args.max_eval_samples ) __snake_case = eval_dataset.select(range(_UpperCAmelCase ) ) with training_args.main_process_first(desc="validation dataset map pre-processing" ): __snake_case = eval_dataset.map( _UpperCAmelCase , batched=_UpperCAmelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator __snake_case = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=_UpperCAmelCase , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(_UpperCAmelCase : Dict ): __snake_case , __snake_case = eval_predictions __snake_case = np.argmax(_UpperCAmelCase , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer __snake_case = Trainer( model=_UpperCAmelCase , args=_UpperCAmelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=_UpperCAmelCase , data_collator=_UpperCAmelCase , compute_metrics=_UpperCAmelCase , ) # Training if training_args.do_train: __snake_case = None if training_args.resume_from_checkpoint is not None: __snake_case = training_args.resume_from_checkpoint elif last_checkpoint is not None: __snake_case = last_checkpoint __snake_case = trainer.train(resume_from_checkpoint=_UpperCAmelCase ) trainer.save_model() # Saves the tokenizer too for easy upload __snake_case = train_result.metrics __snake_case = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_UpperCAmelCase ) ) __snake_case = min(_UpperCAmelCase , len(_UpperCAmelCase ) ) trainer.log_metrics("train" , _UpperCAmelCase ) trainer.save_metrics("train" , _UpperCAmelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("*** Evaluate ***" ) __snake_case = trainer.evaluate() __snake_case = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(_UpperCAmelCase ) __snake_case = min(_UpperCAmelCase , len(_UpperCAmelCase ) ) trainer.log_metrics("eval" , _UpperCAmelCase ) trainer.save_metrics("eval" , _UpperCAmelCase ) __snake_case = { "finetuned_from": model_args.model_name_or_path, "tasks": "multiple-choice", "dataset_tags": "swag", "dataset_args": "regular", "dataset": "SWAG", "language": "en", } if training_args.push_to_hub: trainer.push_to_hub(**_UpperCAmelCase ) else: trainer.create_model_card(**_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : Any ) -> Union[str, Any]: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : Optional[Any] , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __snake_case = input_file.read() __snake_case = regexp.search(a_ ) return match def A ( self : Any , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __snake_case = regexp.finditer(a_ ) __snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A ( self : Optional[int] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a_ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(a_ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
1
'''simple docstring''' from dataclasses import dataclass from typing import Tuple import numpy as np import torch @dataclass class SCREAMING_SNAKE_CASE__ : __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 # [batch_size x 3] __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 __SCREAMING_SNAKE_CASE = 42 def A ( self : List[str] ): """simple docstring""" assert self.x.shape[0] == self.y.shape[0] == self.z.shape[0] == self.origin.shape[0] assert self.x.shape[1] == self.y.shape[1] == self.z.shape[1] == self.origin.shape[1] == 3 assert len(self.x.shape ) == len(self.y.shape ) == len(self.z.shape ) == len(self.origin.shape ) == 2 def A ( self : Dict ): """simple docstring""" return torch.from_numpy(np.array([self.width, self.height] , dtype=np.floataa ) ) def A ( self : Optional[Any] ): """simple docstring""" return torch.from_numpy(np.array([self.x_fov, self.y_fov] , dtype=np.floataa ) ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = torch.arange(self.height * self.width ) __snake_case = torch.stack( [ pixel_indices % self.width, torch.div(a_ , self.width , rounding_mode="trunc" ), ] , axis=1 , ) return coords @property def A ( self : Optional[int] ): """simple docstring""" __snake_case , *__snake_case = self.shape __snake_case = int(np.prod(a_ ) ) __snake_case = self.get_image_coords() __snake_case = torch.broadcast_to(coords.unsqueeze(0 ) , [batch_size * inner_batch_size, *coords.shape] ) __snake_case = self.get_camera_rays(a_ ) __snake_case = rays.view(a_ , inner_batch_size * self.height * self.width , 2 , 3 ) return rays def A ( self : Optional[int] , a_ : torch.Tensor ): """simple docstring""" __snake_case , *__snake_case , __snake_case = coords.shape assert n_coords == 2 assert batch_size == self.origin.shape[0] __snake_case = coords.view(a_ , -1 , 2 ) __snake_case = self.resolution() __snake_case = self.fov() __snake_case = (flat.float() / (res - 1)) * 2 - 1 __snake_case = fracs * torch.tan(fov / 2 ) __snake_case = fracs.view(a_ , -1 , 2 ) __snake_case = ( self.z.view(a_ , 1 , 3 ) + self.x.view(a_ , 1 , 3 ) * fracs[:, :, :1] + self.y.view(a_ , 1 , 3 ) * fracs[:, :, 1:] ) __snake_case = directions / directions.norm(dim=-1 , keepdim=a_ ) __snake_case = torch.stack( [ torch.broadcast_to(self.origin.view(a_ , 1 , 3 ) , [batch_size, directions.shape[1], 3] ), directions, ] , dim=2 , ) return rays.view(a_ , *a_ , 2 , 3 ) def A ( self : List[str] , a_ : int , a_ : int ): """simple docstring""" assert width * self.height == height * self.width, "The aspect ratio should not change." return DifferentiableProjectiveCamera( origin=self.origin , x=self.x , y=self.y , z=self.z , width=a_ , height=a_ , x_fov=self.x_fov , y_fov=self.y_fov , ) def __UpperCAmelCase ( _UpperCAmelCase : int ) -> DifferentiableProjectiveCamera: __snake_case = [] __snake_case = [] __snake_case = [] __snake_case = [] for theta in np.linspace(0 , 2 * np.pi , num=20 ): __snake_case = np.array([np.sin(_UpperCAmelCase ), np.cos(_UpperCAmelCase ), -0.5] ) z /= np.sqrt(np.sum(z**2 ) ) __snake_case = -z * 4 __snake_case = np.array([np.cos(_UpperCAmelCase ), -np.sin(_UpperCAmelCase ), 0.0] ) __snake_case = np.cross(_UpperCAmelCase , _UpperCAmelCase ) origins.append(_UpperCAmelCase ) xs.append(_UpperCAmelCase ) ys.append(_UpperCAmelCase ) zs.append(_UpperCAmelCase ) return DifferentiableProjectiveCamera( origin=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , x=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , y=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , z=torch.from_numpy(np.stack(_UpperCAmelCase , axis=0 ) ).float() , width=_UpperCAmelCase , height=_UpperCAmelCase , x_fov=0.7 , y_fov=0.7 , shape=(1, len(_UpperCAmelCase )) , )
69
'''simple docstring''' a : Dict = range(2, 20 + 1) a : Optional[int] = [10**k for k in range(ks[-1] + 1)] a : dict[int, dict[int, list[list[int]]]] = {} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] ) -> int: __snake_case = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) __snake_case = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) __snake_case , __snake_case = 0, 0 __snake_case = n - i __snake_case = memo.get(_UpperCAmelCase ) if sub_memo is not None: __snake_case = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over __snake_case = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __snake_case = _k break if max_jump >= 0: __snake_case , __snake_case , __snake_case = jumps[max_jump] # since the difference between jumps is cached, add c __snake_case = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: __snake_case = [] else: __snake_case = {c: []} __snake_case = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __snake_case , __snake_case = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __snake_case , __snake_case = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped __snake_case = sub_memo[c] # keep jumps sorted by # of terms skipped __snake_case = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __snake_case = i __snake_case , __snake_case , __snake_case = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __snake_case = ds_c + ds_b diff += addend __snake_case = 0 for j in range(_UpperCAmelCase ): __snake_case = a_i[j] + addend __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Tuple: for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case = digits[j] + addend if s >= 10: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) __snake_case = addend // 10 + quotient else: __snake_case = s __snake_case = addend // 10 if addend == 0: break while addend > 0: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int = 10**15 ) -> int: __snake_case = [1] __snake_case = 1 __snake_case = 0 while True: __snake_case , __snake_case = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break __snake_case = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
69
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : List[str] = {} class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """llama""" __SCREAMING_SNAKE_CASE = ["""past_key_values"""] def __init__( self : Optional[int] , a_ : str=32_000 , a_ : Any=4_096 , a_ : List[Any]=11_008 , a_ : List[str]=32 , a_ : Any=32 , a_ : str=None , a_ : str="silu" , a_ : Optional[int]=2_048 , a_ : Optional[Any]=0.02 , a_ : int=1e-6 , a_ : Any=True , a_ : List[Any]=0 , a_ : int=1 , a_ : List[Any]=2 , a_ : List[Any]=1 , a_ : Union[str, Any]=False , a_ : Optional[int]=None , **a_ : Optional[int] , ): """simple docstring""" __snake_case = vocab_size __snake_case = max_position_embeddings __snake_case = hidden_size __snake_case = intermediate_size __snake_case = num_hidden_layers __snake_case = num_attention_heads # for backward compatibility if num_key_value_heads is None: __snake_case = num_attention_heads __snake_case = num_key_value_heads __snake_case = hidden_act __snake_case = initializer_range __snake_case = rms_norm_eps __snake_case = pretraining_tp __snake_case = use_cache __snake_case = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=a_ , bos_token_id=a_ , eos_token_id=a_ , tie_word_embeddings=a_ , **a_ , ) def A ( self : List[str] ): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , a_ ) or len(self.rope_scaling ) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " f'''got {self.rope_scaling}''' ) __snake_case = self.rope_scaling.get("type" , a_ ) __snake_case = self.rope_scaling.get("factor" , a_ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f'''`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}''' ) if rope_scaling_factor is None or not isinstance(a_ , a_ ) or rope_scaling_factor <= 1.0: raise ValueError(f'''`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}''' )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
1
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Dict ): """simple docstring""" __snake_case = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) __snake_case = get_activation("gelu" ) self.assertTrue(torch.allclose(gelu_python(a_ ) , torch_builtin(a_ ) ) ) self.assertFalse(torch.allclose(gelu_python(a_ ) , gelu_new(a_ ) ) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) __snake_case = get_activation("gelu" ) __snake_case = get_activation("gelu_10" ) __snake_case = torch_builtin(a_ ) __snake_case = geluaa(a_ ) __snake_case = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(a_ ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def A ( self : Tuple ): """simple docstring""" get_activation("gelu" ) get_activation("gelu_10" ) get_activation("gelu_fast" ) get_activation("gelu_new" ) get_activation("gelu_python" ) get_activation("gelu_pytorch_tanh" ) get_activation("linear" ) get_activation("mish" ) get_activation("quick_gelu" ) get_activation("relu" ) get_activation("sigmoid" ) get_activation("silu" ) get_activation("swish" ) get_activation("tanh" ) with self.assertRaises(a_ ): get_activation("bogus" ) with self.assertRaises(a_ ): get_activation(a_ ) def A ( self : str ): """simple docstring""" __snake_case = get_activation("gelu" ) __snake_case = 1 __snake_case = get_activation("gelu" ) self.assertEqual(acta.a , 1 ) with self.assertRaises(a_ ): __snake_case = acta.a
69
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
1
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = ["""image_processor""", """tokenizer"""] __SCREAMING_SNAKE_CASE = """LayoutLMv3ImageProcessor""" __SCREAMING_SNAKE_CASE = ("""LayoutLMv3Tokenizer""", """LayoutLMv3TokenizerFast""") def __init__( self : str , a_ : Dict=None , a_ : Dict=None , **a_ : Optional[int] ): """simple docstring""" __snake_case = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , a_ , ) __snake_case = kwargs.pop("feature_extractor" ) __snake_case = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(a_ , a_ ) def __call__( self : Any , a_ : Union[str, Any] , a_ : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , a_ : Optional[Union[PreTokenizedInput, List[PreTokenizedInput]]] = None , a_ : Union[List[List[int]], List[List[List[int]]]] = None , a_ : Optional[Union[List[int], List[List[int]]]] = None , a_ : bool = True , a_ : Union[bool, str, PaddingStrategy] = False , a_ : Union[bool, str, TruncationStrategy] = None , a_ : Optional[int] = None , a_ : int = 0 , a_ : Optional[int] = None , a_ : Optional[bool] = None , a_ : Optional[bool] = None , a_ : bool = False , a_ : bool = False , a_ : bool = False , a_ : bool = False , a_ : bool = True , a_ : Optional[Union[str, TensorType]] = None , **a_ : Tuple , ): """simple docstring""" if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( "You cannot provide bounding boxes if you initialized the image processor with apply_ocr set to True." ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( "You cannot provide word labels if you initialized the image processor with apply_ocr set to True." ) # first, apply the image processor __snake_case = self.image_processor(images=a_ , return_tensors=a_ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(a_ , a_ ): __snake_case = [text] # add batch dimension (as the image processor always adds a batch dimension) __snake_case = features["words"] __snake_case = self.tokenizer( text=text if text is not None else features["words"] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["boxes"] , word_labels=a_ , add_special_tokens=a_ , padding=a_ , truncation=a_ , max_length=a_ , stride=a_ , pad_to_multiple_of=a_ , return_token_type_ids=a_ , return_attention_mask=a_ , return_overflowing_tokens=a_ , return_special_tokens_mask=a_ , return_offsets_mapping=a_ , return_length=a_ , verbose=a_ , return_tensors=a_ , **a_ , ) # add pixel values __snake_case = features.pop("pixel_values" ) if return_overflowing_tokens is True: __snake_case = self.get_overflowing_images(a_ , encoded_inputs["overflow_to_sample_mapping"] ) __snake_case = images return encoded_inputs def A ( self : List[Any] , a_ : int , a_ : int ): """simple docstring""" __snake_case = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(a_ ) != len(a_ ): raise ValueError( "Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got" f''' {len(a_ )} and {len(a_ )}''' ) return images_with_overflow def A ( self : str , *a_ : str , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : str , *a_ : int , **a_ : Tuple ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ ) @property def A ( self : Dict ): """simple docstring""" return ["input_ids", "bbox", "attention_mask", "pixel_values"] @property def A ( self : int ): """simple docstring""" warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , a_ , ) return self.image_processor_class @property def A ( self : Dict ): """simple docstring""" warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , a_ , ) return self.image_processor
69
'''simple docstring''' import mpmath # for roots of unity import numpy as np class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , a_ : Optional[int]=None , a_ : int=None ): """simple docstring""" __snake_case = list(poly_a or [0] )[:] __snake_case = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() __snake_case = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() __snake_case = len(self.polyB ) # Add 0 to make lengths equal a power of 2 __snake_case = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform __snake_case = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product __snake_case = self.__multiply() def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" __snake_case = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB] # Corner case if len(a_ ) <= 1: return dft[0] # __snake_case = self.c_max_length // 2 while next_ncol > 0: __snake_case = [[] for i in range(a_ )] __snake_case = self.root**next_ncol # First half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update __snake_case = new_dft __snake_case = next_ncol // 2 return dft[0] def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.__dft("A" ) __snake_case = self.__dft("B" ) __snake_case = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT __snake_case = 2 while next_ncol <= self.c_max_length: __snake_case = [[] for i in range(a_ )] __snake_case = self.root ** (next_ncol // 2) __snake_case = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update __snake_case = new_inverse_c next_ncol *= 2 # Unpack __snake_case = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Optional[int] ): """simple docstring""" __snake_case = "A = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) __snake_case = "B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) __snake_case = "A*B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """SpeechT5FeatureExtractor""" __SCREAMING_SNAKE_CASE = """SpeechT5Tokenizer""" def __init__( self : List[Any] , a_ : str , a_ : str ): """simple docstring""" super().__init__(a_ , a_ ) def __call__( self : Dict , *a_ : Tuple , **a_ : List[str] ): """simple docstring""" __snake_case = kwargs.pop("audio" , a_ ) __snake_case = kwargs.pop("text" , a_ ) __snake_case = kwargs.pop("text_target" , a_ ) __snake_case = kwargs.pop("audio_target" , a_ ) __snake_case = kwargs.pop("sampling_rate" , a_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __snake_case = self.feature_extractor(a_ , *a_ , sampling_rate=a_ , **a_ ) elif text is not None: __snake_case = self.tokenizer(a_ , **a_ ) else: __snake_case = None if audio_target is not None: __snake_case = self.feature_extractor(audio_target=a_ , *a_ , sampling_rate=a_ , **a_ ) __snake_case = targets["input_values"] elif text_target is not None: __snake_case = self.tokenizer(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : str , **a_ : Dict ): """simple docstring""" __snake_case = kwargs.pop("input_values" , a_ ) __snake_case = kwargs.pop("input_ids" , a_ ) __snake_case = kwargs.pop("labels" , a_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) elif input_ids is not None: __snake_case = self.tokenizer.pad(a_ , **a_ ) else: __snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(a_ , a_ ) and "input_ids" in labels[0]): __snake_case = self.tokenizer.pad(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = self.feature_extractor.feature_size __snake_case = self.feature_extractor.num_mel_bins __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) __snake_case = feature_size_hack __snake_case = targets["input_values"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : Any , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : Optional[int] , *a_ : Union[str, Any] , **a_ : str ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ )
69
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = KandinskyVaaImgaImgPipeline __SCREAMING_SNAKE_CASE = ["""image_embeds""", """negative_image_embeds""", """image"""] __SCREAMING_SNAKE_CASE = [ """image_embeds""", """negative_image_embeds""", """image""", ] __SCREAMING_SNAKE_CASE = [ """generator""", """height""", """width""", """strength""", """guidance_scale""", """num_inference_steps""", """return_dict""", """guidance_scale""", """num_images_per_prompt""", """output_type""", """return_dict""", ] __SCREAMING_SNAKE_CASE = False @property def A ( self : Any ): """simple docstring""" return 32 @property def A ( self : str ): """simple docstring""" return 32 @property def A ( self : List[Any] ): """simple docstring""" return self.time_input_dim @property def A ( self : Optional[Any] ): """simple docstring""" return self.time_input_dim * 4 @property def A ( self : Tuple ): """simple docstring""" return 100 @property def A ( self : List[str] ): """simple docstring""" torch.manual_seed(0 ) __snake_case = { "in_channels": 4, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } __snake_case = UNetaDConditionModel(**a_ ) return model @property def A ( self : Any ): """simple docstring""" return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def A ( self : Optional[Any] ): """simple docstring""" torch.manual_seed(0 ) __snake_case = VQModel(**self.dummy_movq_kwargs ) return model def A ( self : List[Any] ): """simple docstring""" __snake_case = self.dummy_unet __snake_case = self.dummy_movq __snake_case = { "num_train_timesteps": 1_000, "beta_schedule": "linear", "beta_start": 0.00085, "beta_end": 0.012, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } __snake_case = DDIMScheduler(**a_ ) __snake_case = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def A ( self : str , a_ : Any , a_ : List[Any]=0 ): """simple docstring""" __snake_case = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(a_ ) ).to(a_ ) __snake_case = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( a_ ) # create init_image __snake_case = floats_tensor((1, 3, 64, 64) , rng=random.Random(a_ ) ).to(a_ ) __snake_case = image.cpu().permute(0 , 2 , 3 , 1 )[0] __snake_case = Image.fromarray(np.uinta(a_ ) ).convert("RGB" ).resize((256, 256) ) if str(a_ ).startswith("mps" ): __snake_case = torch.manual_seed(a_ ) else: __snake_case = torch.Generator(device=a_ ).manual_seed(a_ ) __snake_case = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def A ( self : Any ): """simple docstring""" __snake_case = "cpu" __snake_case = self.get_dummy_components() __snake_case = self.pipeline_class(**a_ ) __snake_case = pipe.to(a_ ) pipe.set_progress_bar_config(disable=a_ ) __snake_case = pipe(**self.get_dummy_inputs(a_ ) ) __snake_case = output.images __snake_case = pipe( **self.get_dummy_inputs(a_ ) , return_dict=a_ , )[0] __snake_case = image[0, -3:, -3:, -1] __snake_case = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __snake_case = np.array( [0.6199778, 0.63984406, 0.46145785, 0.62944984, 0.5622215, 0.47306132, 0.47441456, 0.4607606, 0.48719263] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Any ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : Tuple ): """simple docstring""" __snake_case = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_img2img_frog.npy" ) __snake_case = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) __snake_case = "A red cartoon frog, 4k" __snake_case = KandinskyVaaPriorPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" , torch_dtype=torch.floataa ) pipe_prior.to(a_ ) __snake_case = KandinskyVaaImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-decoder" , torch_dtype=torch.floataa ) __snake_case = pipeline.to(a_ ) pipeline.set_progress_bar_config(disable=a_ ) __snake_case = torch.Generator(device="cpu" ).manual_seed(0 ) __snake_case , __snake_case = pipe_prior( a_ , generator=a_ , num_inference_steps=5 , negative_prompt="" , ).to_tuple() __snake_case = pipeline( image=a_ , image_embeds=a_ , negative_image_embeds=a_ , generator=a_ , num_inference_steps=100 , height=768 , width=768 , strength=0.2 , output_type="np" , ) __snake_case = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(a_ , a_ )
69
'''simple docstring''' import json import os import torch from diffusers import UNetaDModel os.makedirs('''hub/hopper-medium-v2/unet/hor32''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/unet/hor128''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/value_function''', exist_ok=True) def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> str: if hor == 1_28: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D") elif hor == 32: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 64, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D", "UpResnetBlock1D") __snake_case = torch.load(F'''/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch''' ) __snake_case = model.state_dict() __snake_case = { "down_block_types": down_block_types, "block_out_channels": block_out_channels, "up_block_types": up_block_types, "layers_per_block": 1, "use_timestep_embedding": True, "out_block_type": "OutConv1DBlock", "norm_num_groups": 8, "downsample_each_block": False, "in_channels": 14, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "flip_sin_to_cos": False, "freq_shift": 1, "sample_size": 6_55_36, "mid_block_type": "MidResTemporalBlock1D", "act_fn": "mish", } __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , F'''hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin''' ) with open(F'''hub/hopper-medium-v2/unet/hor{hor}/config.json''' , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) def __UpperCAmelCase ( ) -> List[Any]: __snake_case = { "in_channels": 14, "down_block_types": ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D"), "up_block_types": (), "out_block_type": "ValueFunction", "mid_block_type": "ValueFunctionMidBlock1D", "block_out_channels": (32, 64, 1_28, 2_56), "layers_per_block": 1, "downsample_each_block": True, "sample_size": 6_55_36, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "use_timestep_embedding": True, "flip_sin_to_cos": False, "freq_shift": 1, "norm_num_groups": 8, "act_fn": "mish", } __snake_case = torch.load("/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch" ) __snake_case = model __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , "hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin" ) with open("hub/hopper-medium-v2/value_function/config.json" , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": unet(32) # unet(128) value_function()
69
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a : Tuple = logging.get_logger(__name__) a : Tuple = {'''openai-gpt''': '''https://huggingface.co/openai-gpt/resolve/main/config.json'''} class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """openai-gpt""" __SCREAMING_SNAKE_CASE = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : Optional[int] , a_ : int=40_478 , a_ : List[str]=512 , a_ : Any=768 , a_ : Dict=12 , a_ : Tuple=12 , a_ : List[str]="gelu" , a_ : Optional[int]=0.1 , a_ : Tuple=0.1 , a_ : List[Any]=0.1 , a_ : Optional[Any]=1e-5 , a_ : int=0.02 , a_ : str="cls_index" , a_ : Tuple=True , a_ : Dict=None , a_ : List[Any]=True , a_ : Optional[int]=0.1 , **a_ : Tuple , ): """simple docstring""" __snake_case = vocab_size __snake_case = n_positions __snake_case = n_embd __snake_case = n_layer __snake_case = n_head __snake_case = afn __snake_case = resid_pdrop __snake_case = embd_pdrop __snake_case = attn_pdrop __snake_case = layer_norm_epsilon __snake_case = initializer_range __snake_case = summary_type __snake_case = summary_use_proj __snake_case = summary_activation __snake_case = summary_first_dropout __snake_case = summary_proj_to_labels super().__init__(**a_ )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 1_00_00_00 ) -> int: __snake_case = 1 __snake_case = 1 __snake_case = {1: 1} for inputa in range(2 , _UpperCAmelCase ): __snake_case = 0 __snake_case = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __snake_case = (3 * number) + 1 counter += 1 if inputa not in counters: __snake_case = counter if counter > pre_counter: __snake_case = inputa __snake_case = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
69
1
import random def __lowercase ( snake_case, snake_case, snake_case ): """simple docstring""" __magic_name__ :Tuple = a[left_index] __magic_name__ :str = left_index + 1 for j in range(left_index + 1, snake_case ): if a[j] < pivot: __magic_name__ , __magic_name__ :Any = a[i], a[j] i += 1 __magic_name__ , __magic_name__ :int = a[i - 1], a[left_index] return i - 1 def __lowercase ( snake_case, snake_case, snake_case ): """simple docstring""" if left < right: __magic_name__ :Optional[Any] = random.randint(snake_case, right - 1 ) __magic_name__ , __magic_name__ :Tuple = ( a[left], a[pivot], ) # switches the pivot with the left most bound __magic_name__ :Tuple = partition(snake_case, snake_case, snake_case ) quick_sort_random( snake_case, snake_case, snake_case ) # recursive quicksort to the left of the pivot point quick_sort_random( snake_case, pivot_index + 1, snake_case ) # recursive quicksort to the right of the pivot point def __lowercase ( ): """simple docstring""" __magic_name__ :Tuple = input('''Enter numbers separated by a comma:\n''' ).strip() __magic_name__ :int = [int(snake_case ) for item in user_input.split(''',''' )] quick_sort_random(snake_case, 0, len(snake_case ) ) print(snake_case ) if __name__ == "__main__": main()
0
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """SpeechT5FeatureExtractor""" __SCREAMING_SNAKE_CASE = """SpeechT5Tokenizer""" def __init__( self : List[Any] , a_ : str , a_ : str ): """simple docstring""" super().__init__(a_ , a_ ) def __call__( self : Dict , *a_ : Tuple , **a_ : List[str] ): """simple docstring""" __snake_case = kwargs.pop("audio" , a_ ) __snake_case = kwargs.pop("text" , a_ ) __snake_case = kwargs.pop("text_target" , a_ ) __snake_case = kwargs.pop("audio_target" , a_ ) __snake_case = kwargs.pop("sampling_rate" , a_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __snake_case = self.feature_extractor(a_ , *a_ , sampling_rate=a_ , **a_ ) elif text is not None: __snake_case = self.tokenizer(a_ , **a_ ) else: __snake_case = None if audio_target is not None: __snake_case = self.feature_extractor(audio_target=a_ , *a_ , sampling_rate=a_ , **a_ ) __snake_case = targets["input_values"] elif text_target is not None: __snake_case = self.tokenizer(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : str , **a_ : Dict ): """simple docstring""" __snake_case = kwargs.pop("input_values" , a_ ) __snake_case = kwargs.pop("input_ids" , a_ ) __snake_case = kwargs.pop("labels" , a_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) elif input_ids is not None: __snake_case = self.tokenizer.pad(a_ , **a_ ) else: __snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(a_ , a_ ) and "input_ids" in labels[0]): __snake_case = self.tokenizer.pad(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = self.feature_extractor.feature_size __snake_case = self.feature_extractor.num_mel_bins __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) __snake_case = feature_size_hack __snake_case = targets["input_values"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : Any , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : Optional[int] , *a_ : Union[str, Any] , **a_ : str ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ )
69
0
import itertools import random import unittest import numpy as np from transformers import ASTFeatureExtractor from transformers.testing_utils import require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin __snake_case = random.Random() if is_torch_available(): import torch def _A ( _lowercase , _lowercase=1.0 , _lowercase=None , _lowercase=None ) -> Dict: """simple docstring""" if rng is None: __UpperCamelCase = global_rng __UpperCamelCase = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values class __lowerCamelCase (unittest.TestCase ): def __init__( self: List[Any],A_: int,A_: Optional[int]=7,A_: Tuple=400,A_: Optional[int]=2000,A_: str=1,A_: Dict=0.0,A_: Any=1_6000,A_: List[Any]=True,A_: List[Any]=True,): '''simple docstring''' __UpperCamelCase = parent __UpperCamelCase = batch_size __UpperCamelCase = min_seq_length __UpperCamelCase = max_seq_length __UpperCamelCase = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __UpperCamelCase = feature_size __UpperCamelCase = padding_value __UpperCamelCase = sampling_rate __UpperCamelCase = return_attention_mask __UpperCamelCase = do_normalize def snake_case_ ( self: int ): '''simple docstring''' return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def snake_case_ ( self: Any,A_: Tuple=False,A_: int=False ): '''simple docstring''' def _flatten(A_: Optional[int] ): return list(itertools.chain(*A_ ) ) if equal_length: __UpperCamelCase = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size __UpperCamelCase = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length,self.max_seq_length,self.seq_length_diff ) ] if numpify: __UpperCamelCase = [np.asarray(A_ ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __lowerCamelCase (_a , unittest.TestCase ): _lowercase = ASTFeatureExtractor def snake_case_ ( self: Optional[Any] ): '''simple docstring''' __UpperCamelCase = ASTFeatureExtractionTester(self ) def snake_case_ ( self: Optional[Any] ): '''simple docstring''' __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __UpperCamelCase = [floats_list((1, x) )[0] for x in range(800,1400,200 )] __UpperCamelCase = [np.asarray(A_ ) for speech_input in speech_inputs] # Test not batched input __UpperCamelCase = feat_extract(speech_inputs[0],return_tensors='np' ).input_values __UpperCamelCase = feat_extract(np_speech_inputs[0],return_tensors='np' ).input_values self.assertTrue(np.allclose(A_,A_,atol=1E-3 ) ) # Test batched __UpperCamelCase = feat_extract(A_,padding=A_,return_tensors='np' ).input_values __UpperCamelCase = feat_extract(A_,padding=A_,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(A_,A_ ): self.assertTrue(np.allclose(A_,A_,atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __UpperCamelCase = [floats_list((1, x) )[0] for x in (800, 800, 800)] __UpperCamelCase = np.asarray(A_ ) __UpperCamelCase = feat_extract(A_,return_tensors='np' ).input_values __UpperCamelCase = feat_extract(A_,return_tensors='np' ).input_values for enc_seq_a, enc_seq_a in zip(A_,A_ ): self.assertTrue(np.allclose(A_,A_,atol=1E-3 ) ) @require_torch def snake_case_ ( self: Optional[Any] ): '''simple docstring''' import torch __UpperCamelCase = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __UpperCamelCase = np.random.rand(100 ).astype(np.floataa ) __UpperCamelCase = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __UpperCamelCase = feature_extractor.pad([{'input_values': inputs}],return_tensors='np' ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) __UpperCamelCase = feature_extractor.pad([{'input_values': inputs}],return_tensors='pt' ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def snake_case_ ( self: Any,A_: Union[str, Any] ): '''simple docstring''' from datasets import load_dataset __UpperCamelCase = load_dataset('hf-internal-testing/librispeech_asr_dummy','clean',split='validation' ) # automatic decoding with librispeech __UpperCamelCase = ds.sort('id' ).select(range(A_ ) )[:num_samples]['audio'] return [x["array"] for x in speech_samples] @require_torch def snake_case_ ( self: Union[str, Any] ): '''simple docstring''' __UpperCamelCase = torch.tensor( [-0.9_8_9_4, -1.2_7_7_6, -0.9_0_6_6, -1.2_7_7_6, -0.9_3_4_9, -1.2_6_0_9, -1.0_3_8_6, -1.2_7_7_6, -1.1_5_6_1, -1.2_7_7_6, -1.2_0_5_2, -1.2_7_2_3, -1.2_1_9_0, -1.2_1_3_2, -1.2_7_7_6, -1.1_1_3_3, -1.1_9_5_3, -1.1_3_4_3, -1.1_5_8_4, -1.2_2_0_3, -1.1_7_7_0, -1.2_4_7_4, -1.2_3_8_1, -1.1_9_3_6, -0.9_2_7_0, -0.8_3_1_7, -0.8_0_4_9, -0.7_7_0_6, -0.7_5_6_5, -0.7_8_6_9] ) # fmt: on __UpperCamelCase = self._load_datasamples(1 ) __UpperCamelCase = ASTFeatureExtractor() __UpperCamelCase = feature_extractor(A_,return_tensors='pt' ).input_values self.assertEquals(input_values.shape,(1, 1024, 128) ) self.assertTrue(torch.allclose(input_values[0, 0, :30],A_,atol=1E-4 ) )
1
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : Optional[Any] , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __snake_case = input_file.read() __snake_case = regexp.search(a_ ) return match def A ( self : Any , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __snake_case = regexp.finditer(a_ ) __snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A ( self : Optional[int] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a_ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(a_ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
0
import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class lowerCamelCase__ ( unittest.TestCase): """simple docstring""" def __init__( self : Union[str, Any] , __lowerCAmelCase : Dict , __lowerCAmelCase : List[str]=7 , __lowerCAmelCase : int=3 , __lowerCAmelCase : int=18 , __lowerCAmelCase : Tuple=30 , __lowerCAmelCase : List[str]=4_00 , __lowerCAmelCase : List[Any]=True , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : Tuple=None , __lowerCAmelCase : Dict=True , __lowerCAmelCase : Dict=[0.5, 0.5, 0.5] , __lowerCAmelCase : Optional[Any]=[0.5, 0.5, 0.5] , __lowerCAmelCase : Tuple=False , ) -> List[str]: _A = size if size is not None else {'''height''': 20, '''width''': 20} _A = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} _A = parent _A = batch_size _A = num_channels _A = image_size _A = min_resolution _A = max_resolution _A = do_resize _A = size _A = do_center_crop _A = crop_size _A = do_normalize _A = image_mean _A = image_std _A = do_reduce_labels def snake_case_ ( self : Optional[int] ) -> Tuple: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def SCREAMING_SNAKE_CASE_ ( ) -> Tuple: _A = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) _A = Image.open(dataset[0]['''file'''] ) _A = Image.open(dataset[1]['''file'''] ) return image, map def SCREAMING_SNAKE_CASE_ ( ) -> Union[str, Any]: _A = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) _A = Image.open(ds[0]['''file'''] ) _A = Image.open(ds[1]['''file'''] ) _A = Image.open(ds[2]['''file'''] ) _A = Image.open(ds[3]['''file'''] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class lowerCamelCase__ ( _A , unittest.TestCase): """simple docstring""" a__ : List[Any] = BeitImageProcessor if is_vision_available() else None def snake_case_ ( self : Optional[Any] ) -> Optional[Any]: _A = BeitImageProcessingTester(self ) @property def snake_case_ ( self : Dict ) -> Optional[int]: return self.image_processor_tester.prepare_image_processor_dict() def snake_case_ ( self : int ) -> List[str]: _A = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__lowerCAmelCase , '''do_resize''' ) ) self.assertTrue(hasattr(__lowerCAmelCase , '''size''' ) ) self.assertTrue(hasattr(__lowerCAmelCase , '''do_center_crop''' ) ) self.assertTrue(hasattr(__lowerCAmelCase , '''center_crop''' ) ) self.assertTrue(hasattr(__lowerCAmelCase , '''do_normalize''' ) ) self.assertTrue(hasattr(__lowerCAmelCase , '''image_mean''' ) ) self.assertTrue(hasattr(__lowerCAmelCase , '''image_std''' ) ) def snake_case_ ( self : int ) -> List[str]: _A = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 20, '''width''': 20} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) self.assertEqual(image_processor.do_reduce_labels , __lowerCAmelCase ) _A = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=__lowerCAmelCase ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) self.assertEqual(image_processor.do_reduce_labels , __lowerCAmelCase ) def snake_case_ ( self : Union[str, Any] ) -> Optional[Any]: pass def snake_case_ ( self : Union[str, Any] ) -> int: # Initialize image_processing _A = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _A = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase , Image.Image ) # Test not batched input _A = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _A = image_processing(__lowerCAmelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def snake_case_ ( self : Optional[Any] ) -> Any: # Initialize image_processing _A = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _A = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCAmelCase , numpify=__lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase , np.ndarray ) # Test not batched input _A = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _A = image_processing(__lowerCAmelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def snake_case_ ( self : List[Any] ) -> Union[str, Any]: # Initialize image_processing _A = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _A = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCAmelCase , torchify=__lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(__lowerCAmelCase , torch.Tensor ) # Test not batched input _A = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched _A = image_processing(__lowerCAmelCase , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def snake_case_ ( self : int ) -> str: # Initialize image_processing _A = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _A = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowerCAmelCase , torchify=__lowerCAmelCase ) _A = [] for image in image_inputs: self.assertIsInstance(__lowerCAmelCase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input _A = image_processing(image_inputs[0] , maps[0] , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( 1, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 2_55 ) # Test batched _A = image_processing(__lowerCAmelCase , __lowerCAmelCase , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 2_55 ) # Test not batched input (PIL images) _A , _A = prepare_semantic_single_inputs() _A = image_processing(__lowerCAmelCase , __lowerCAmelCase , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( 1, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 2_55 ) # Test batched input (PIL images) _A , _A = prepare_semantic_batch_inputs() _A = image_processing(__lowerCAmelCase , __lowerCAmelCase , return_tensors='''pt''' ) self.assertEqual( encoding['''pixel_values'''].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual( encoding['''labels'''].shape , ( 2, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) self.assertEqual(encoding['''labels'''].dtype , torch.long ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 2_55 ) def snake_case_ ( self : List[str] ) -> Dict: # Initialize image_processing _A = self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 _A , _A = prepare_semantic_single_inputs() _A = image_processing(__lowerCAmelCase , __lowerCAmelCase , return_tensors='''pt''' ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 1_50 ) _A = True _A = image_processing(__lowerCAmelCase , __lowerCAmelCase , return_tensors='''pt''' ) self.assertTrue(encoding['''labels'''].min().item() >= 0 ) self.assertTrue(encoding['''labels'''].max().item() <= 2_55 )
2
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : Dict = {'''vocab_file''': '''sentencepiece.model'''} a : Tuple = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } a : str = { '''google/rembert''': 256, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[Any] , a_ : int , a_ : Any=False , a_ : List[Any]=True , a_ : List[Any]=True , a_ : List[Any]="[CLS]" , a_ : List[Any]="[SEP]" , a_ : List[Any]="[UNK]" , a_ : str="[SEP]" , a_ : List[str]="[PAD]" , a_ : Optional[int]="[CLS]" , a_ : List[str]="[MASK]" , **a_ : str , ): """simple docstring""" super().__init__( do_lower_case=a_ , remove_space=a_ , keep_accents=a_ , bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , pad_token=a_ , cls_token=a_ , mask_token=a_ , **a_ , ) __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(a_ ) @property def A ( self : Optional[Any] ): """simple docstring""" return len(self.sp_model ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ): """simple docstring""" __snake_case = self.__dict__.copy() __snake_case = None return state def __setstate__( self : str , a_ : Optional[int] ): """simple docstring""" __snake_case = d __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def A ( self : Tuple , a_ : Optional[int] , a_ : int=False ): """simple docstring""" __snake_case = self.sp_model.EncodeAsPieces(a_ ) return pieces def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" return self.sp_model.PieceToId(a_ ) def A ( self : Optional[Any] , a_ : List[str] ): """simple docstring""" return self.sp_model.IdToPiece(a_ ) def A ( self : Optional[Any] , a_ : int ): """simple docstring""" __snake_case = self.sp_model.decode_pieces(a_ ) return out_string def A ( self : Union[str, Any] , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a_ )) + [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1] def A ( self : Tuple , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A ( self : List[Any] , a_ : str , a_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a_ ): logger.error("Vocabulary path ({}) should be a directory".format(a_ ) ) return __snake_case = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ): copyfile(self.vocab_file , a_ ) return (out_vocab_file,)
69
0
'''simple docstring''' from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) lowerCAmelCase : Dict = _symbol_database.Default() lowerCAmelCase : Union[str, Any] = _descriptor_pool.Default().AddSerializedFile( B'\n\x19sentencepiece_model.proto\x12\rsentencepiece"\x80\x0c\n\x0bTrainerSpec\x12\r\n\x05input\x18\x01 \x03(\t\x12\x14\n\x0cinput_format\x18\x07 \x01(\t\x12\x14\n\x0cmodel_prefix\x18\x02 \x01(\t\x12\x41\n\nmodel_type\x18\x03 \x01(\x0e\x32$.sentencepiece.TrainerSpec.ModelType:\x07UNIGRAM\x12\x18\n\nvocab_size\x18\x04 \x01(\x05:\x04\x38\x30\x30\x30\x12\x17\n\x0f\x61\x63\x63\x65pt_language\x18\x05 \x03(\t\x12 \n\x15self_test_sample_size\x18\x06 \x01(\x05:\x01\x30\x12*\n\x1b\x65nable_differential_privacy\x18\x32 \x01(\x08:\x05\x66\x61lse\x12+\n differential_privacy_noise_level\x18\x33 \x01(\x02:\x01\x30\x12\x32\n\'differential_privacy_clipping_threshold\x18\x34 \x01(\x04:\x01\x30\x12"\n\x12\x63haracter_coverage\x18\n \x01(\x02:\x06\x30.9995\x12\x1e\n\x13input_sentence_size\x18\x0b \x01(\x04:\x01\x30\x12$\n\x16shuffle_input_sentence\x18\x13 \x01(\x08:\x04true\x12 \n\x14mining_sentence_size\x18\x0c \x01(\x05\x42\x02\x18\x01\x12"\n\x16training_sentence_size\x18\r \x01(\x05\x42\x02\x18\x01\x12(\n\x17seed_sentencepiece_size\x18\x0e \x01(\x05:\x07\x31\x30\x30\x30\x30\x30\x30\x12\x1e\n\x10shrinking_factor\x18\x0f \x01(\x02:\x04\x30.75\x12!\n\x13max_sentence_length\x18\x12 \x01(\x05:\x04\x34\x31\x39\x32\x12\x17\n\x0bnum_threads\x18\x10 \x01(\x05:\x02\x31\x36\x12\x1d\n\x12num_sub_iterations\x18\x11 \x01(\x05:\x01\x32\x12$\n\x18max_sentencepiece_length\x18\x14 \x01(\x05:\x02\x31\x36\x12%\n\x17split_by_unicode_script\x18\x15 \x01(\x08:\x04true\x12\x1d\n\x0fsplit_by_number\x18\x17 \x01(\x08:\x04true\x12!\n\x13split_by_whitespace\x18\x16 \x01(\x08:\x04true\x12)\n\x1atreat_whitespace_as_suffix\x18\x18 \x01(\x08:\x05\x66\x61lse\x12+\n\x1c\x61llow_whitespace_only_pieces\x18\x1a \x01(\x08:\x05\x66\x61lse\x12\x1b\n\x0csplit_digits\x18\x19 \x01(\x08:\x05\x66\x61lse\x12#\n\x19pretokenization_delimiter\x18\x35 \x01(\t:\x00\x12\x17\n\x0f\x63ontrol_symbols\x18\x1e \x03(\t\x12\x1c\n\x14user_defined_symbols\x18\x1f \x03(\t\x12\x16\n\x0erequired_chars\x18$ \x01(\t\x12\x1c\n\rbyte_fallback\x18# \x01(\x08:\x05\x66\x61lse\x12+\n\x1dvocabulary_output_piece_score\x18 \x01(\x08:\x04true\x12\x1e\n\x10hard_vocab_limit\x18! \x01(\x08:\x04true\x12\x1c\n\ruse_all_vocab\x18" \x01(\x08:\x05\x66\x61lse\x12\x11\n\x06unk_id\x18( \x01(\x05:\x01\x30\x12\x11\n\x06\x62os_id\x18) \x01(\x05:\x01\x31\x12\x11\n\x06\x65os_id\x18* \x01(\x05:\x01\x32\x12\x12\n\x06pad_id\x18+ \x01(\x05:\x02-1\x12\x18\n\tunk_piece\x18- \x01(\t:\x05<unk>\x12\x16\n\tbos_piece\x18. \x01(\t:\x03<s>\x12\x17\n\teos_piece\x18/ \x01(\t:\x04</s>\x12\x18\n\tpad_piece\x18\x30 \x01(\t:\x05<pad>\x12\x1a\n\x0bunk_surface\x18, \x01(\t:\x05 \xe2\x81\x87 \x12+\n\x1ctrain_extremely_large_corpus\x18\x31 \x01(\x08:\x05\x66\x61lse"5\n\tModelType\x12\x0b\n\x07UNIGRAM\x10\x01\x12\x07\n\x03\x42PE\x10\x02\x12\x08\n\x04WORD\x10\x03\x12\x08\n\x04\x43HAR\x10\x04*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"\xd1\x01\n\x0eNormalizerSpec\x12\x0c\n\x04name\x18\x01 \x01(\t\x12\x1c\n\x14precompiled_charsmap\x18\x02 \x01(\x0c\x12\x1e\n\x10\x61\x64\x64_dummy_prefix\x18\x03 \x01(\x08:\x04true\x12&\n\x18remove_extra_whitespaces\x18\x04 \x01(\x08:\x04true\x12 \n\x12\x65scape_whitespaces\x18\x05 \x01(\x08:\x04true\x12\x1e\n\x16normalization_rule_tsv\x18\x06 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"y\n\x0cSelfTestData\x12\x33\n\x07samples\x18\x01 \x03(\x0b\x32".sentencepiece.SelfTestData.Sample\x1a)\n\x06Sample\x12\r\n\x05input\x18\x01 \x01(\t\x12\x10\n\x08\x65xpected\x18\x02 \x01(\t*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02"\xfe\x03\n\nModelProto\x12\x37\n\x06pieces\x18\x01 \x03(\x0b\x32\'.sentencepiece.ModelProto.SentencePiece\x12\x30\n\x0ctrainer_spec\x18\x02 \x01(\x0b\x32\x1a.sentencepiece.TrainerSpec\x12\x36\n\x0fnormalizer_spec\x18\x03 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x12\x33\n\x0eself_test_data\x18\x04 \x01(\x0b\x32\x1b.sentencepiece.SelfTestData\x12\x38\n\x11\x64\x65normalizer_spec\x18\x05 \x01(\x0b\x32\x1d.sentencepiece.NormalizerSpec\x1a\xd2\x01\n\rSentencePiece\x12\r\n\x05piece\x18\x01 \x01(\t\x12\r\n\x05score\x18\x02 \x01(\x02\x12\x42\n\x04type\x18\x03 \x01(\x0e\x32,.sentencepiece.ModelProto.SentencePiece.Type:\x06NORMAL"T\n\x04Type\x12\n\n\x06NORMAL\x10\x01\x12\x0b\n\x07UNKNOWN\x10\x02\x12\x0b\n\x07\x43ONTROL\x10\x03\x12\x10\n\x0cUSER_DEFINED\x10\x04\x12\x08\n\x04\x42YTE\x10\x06\x12\n\n\x06UNUSED\x10\x05*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02*\t\x08\xc8\x01\x10\x80\x80\x80\x80\x02\x42\x02H\x03' ) lowerCAmelCase : Optional[Any] = globals() _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals) _builder.BuildTopDescriptorsAndMessages(DESCRIPTOR, 'sentencepiece_model_pb2', _globals) if _descriptor._USE_C_DESCRIPTORS is False: lowerCAmelCase : Optional[int] = None lowerCAmelCase : str = B'H\003' # (generated by protobuf compiler, but `_TRAINERSPEC` is not defined) # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["mining_sentence_size"]._serialized_options = b"\030\001" # _TRAINERSPEC.fields_by_name["training_sentence_size"]._options = None # _TRAINERSPEC.fields_by_name["training_sentence_size"]._serialized_options = b"\030\001" lowerCAmelCase : Optional[int] = 45 lowerCAmelCase : Tuple = 15_81 lowerCAmelCase : Tuple = 15_17 lowerCAmelCase : Tuple = 15_70 lowerCAmelCase : Union[str, Any] = 15_84 lowerCAmelCase : Optional[int] = 17_93 lowerCAmelCase : int = 17_95 lowerCAmelCase : Dict = 19_16 lowerCAmelCase : List[Any] = 18_64 lowerCAmelCase : Any = 19_05 lowerCAmelCase : Any = 19_19 lowerCAmelCase : str = 24_29 lowerCAmelCase : str = 22_08 lowerCAmelCase : Any = 24_18 lowerCAmelCase : Dict = 23_23 lowerCAmelCase : Optional[int] = 24_07 # @@protoc_insertion_point(module_scope)
3
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ): """simple docstring""" try: __snake_case = tempfile.mktemp() with open(a_ , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , a_ ) __snake_case = AlbertTokenizer.from_pretrained(a_ ) finally: os.remove(a_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , a_ ) __snake_case = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A ( self : str ): """simple docstring""" __snake_case = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : List[Any] ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a_ , repo_id="test-tokenizer" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a_ , repo_id="valid_org/test-tokenizer-org" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A ( self : List[str] ): """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizerFast.from_pretrained(a_ ) bert_tokenizer.save_pretrained(a_ ) __snake_case = CustomTokenizerFast.from_pretrained(a_ ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a_ , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A ( self : str ): """simple docstring""" __snake_case = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def A ( self : List[Any] ): """simple docstring""" __snake_case = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : str ): """simple docstring""" __snake_case = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def A ( self : Tuple ): """simple docstring""" __snake_case = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def A ( self : Any ): """simple docstring""" __snake_case = Trie() __snake_case = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a_ , ["AB", "C"] )
69
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : List[Any] ): lowerCAmelCase = args.pruning_method lowerCAmelCase = args.threshold lowerCAmelCase = args.model_name_or_path.rstrip('/' ) lowerCAmelCase = args.target_model_path print(F'Load fine-pruned model from {model_name_or_path}' ) lowerCAmelCase = torch.load(os.path.join(_UpperCAmelCase , 'pytorch_model.bin' ) ) lowerCAmelCase = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: lowerCAmelCase = tensor print(F'Copied layer {name}' ) elif "classifier" in name or "qa_output" in name: lowerCAmelCase = tensor print(F'Copied layer {name}' ) elif "bias" in name: lowerCAmelCase = tensor print(F'Copied layer {name}' ) else: if pruning_method == "magnitude": lowerCAmelCase = MagnitudeBinarizer.apply(inputs=_UpperCAmelCase , threshold=_UpperCAmelCase ) lowerCAmelCase = tensor * mask print(F'Pruned layer {name}' ) elif pruning_method == "topK": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F'{prefix_}mask_scores'] lowerCAmelCase = TopKBinarizer.apply(_UpperCAmelCase , _UpperCAmelCase ) lowerCAmelCase = tensor * mask print(F'Pruned layer {name}' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F'{prefix_}mask_scores'] lowerCAmelCase = ThresholdBinarizer.apply(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) lowerCAmelCase = tensor * mask print(F'Pruned layer {name}' ) elif pruning_method == "l0": if "mask_scores" in name: continue lowerCAmelCase = name[:-6] lowerCAmelCase = model[F'{prefix_}mask_scores'] lowerCAmelCase ,lowerCAmelCase = -0.1, 1.1 lowerCAmelCase = torch.sigmoid(_UpperCAmelCase ) lowerCAmelCase = s * (r - l) + l lowerCAmelCase = s_bar.clamp(min=0.0 , max=1.0 ) lowerCAmelCase = tensor * mask print(F'Pruned layer {name}' ) else: raise ValueError('Unknown pruning method' ) if target_model_path is None: lowerCAmelCase = os.path.join( os.path.dirname(_UpperCAmelCase ) , F'bertarized_{os.path.basename(_UpperCAmelCase )}' ) if not os.path.isdir(_UpperCAmelCase ): shutil.copytree(_UpperCAmelCase , _UpperCAmelCase ) print(F'\nCreated folder {target_model_path}' ) torch.save(_UpperCAmelCase , os.path.join(_UpperCAmelCase , 'pytorch_model.bin' ) ) print('\nPruned model saved! See you later!' ) if __name__ == "__main__": __UpperCamelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '''--pruning_method''', choices=['''l0''', '''magnitude''', '''topK''', '''sigmoied_threshold'''], type=str, required=True, help=( '''Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,''' ''' sigmoied_threshold = Soft movement pruning)''' ), ) parser.add_argument( '''--threshold''', type=float, required=False, help=( '''For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.''' '''For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.''' '''Not needed for `l0`''' ), ) parser.add_argument( '''--model_name_or_path''', type=str, required=True, help='''Folder containing the model that was previously fine-pruned''', ) parser.add_argument( '''--target_model_path''', default=None, type=str, required=False, help='''Folder containing the model that was previously fine-pruned''', ) __UpperCamelCase : List[Any] = parser.parse_args() main(args)
4
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 __snake_case , __snake_case = 1, 1 for _ in range(number_of_steps - 1 ): __snake_case , __snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
0
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax _lowercase = logging.get_logger(__name__) @add_end_docstrings(_SCREAMING_SNAKE_CASE ) class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , **_lowercase ): """simple docstring""" super().__init__(**_lowercase ) requires_backends(self , """vision""" ) self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == """tf""" else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING ) def __call__( self , _lowercase , **_lowercase ): """simple docstring""" return super().__call__(_lowercase , **_lowercase ) def _lowercase ( self , **_lowercase ): """simple docstring""" _lowerCAmelCase = {} if "candidate_labels" in kwargs: _lowerCAmelCase = kwargs["""candidate_labels"""] if "hypothesis_template" in kwargs: _lowerCAmelCase = kwargs["""hypothesis_template"""] return preprocess_params, {}, {} def _lowercase ( self , _lowercase , _lowercase=None , _lowercase="This is a photo of {}." ): """simple docstring""" _lowerCAmelCase = load_image(_lowercase ) _lowerCAmelCase = self.image_processor(images=[image] , return_tensors=self.framework ) _lowerCAmelCase = candidate_labels _lowerCAmelCase = [hypothesis_template.format(_lowercase ) for x in candidate_labels] _lowerCAmelCase = self.tokenizer(_lowercase , return_tensors=self.framework , padding=_lowercase ) _lowerCAmelCase = [text_inputs] return inputs def _lowercase ( self , _lowercase ): """simple docstring""" _lowerCAmelCase = model_inputs.pop("""candidate_labels""" ) _lowerCAmelCase = model_inputs.pop("""text_inputs""" ) if isinstance(text_inputs[0] , _lowercase ): _lowerCAmelCase = text_inputs[0] else: # Batching case. _lowerCAmelCase = text_inputs[0][0] _lowerCAmelCase = self.model(**_lowercase , **_lowercase ) _lowerCAmelCase = { """candidate_labels""": candidate_labels, """logits""": outputs.logits_per_image, } return model_outputs def _lowercase ( self , _lowercase ): """simple docstring""" _lowerCAmelCase = model_outputs.pop("""candidate_labels""" ) _lowerCAmelCase = model_outputs["""logits"""][0] if self.framework == "pt": _lowerCAmelCase = logits.softmax(dim=-1 ).squeeze(-1 ) _lowerCAmelCase = probs.tolist() if not isinstance(_lowercase , _lowercase ): _lowerCAmelCase = [scores] elif self.framework == "tf": _lowerCAmelCase = stable_softmax(_lowercase , axis=-1 ) _lowerCAmelCase = probs.numpy().tolist() else: raise ValueError(F'Unsupported framework: {self.framework}' ) _lowerCAmelCase = [ {"""score""": score, """label""": candidate_label} for score, candidate_label in sorted(zip(_lowercase , _lowercase ) , key=lambda _lowercase : -x[0] ) ] return result
5
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str ) -> str: return " ".join( "".join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
69
0
import requests from bsa import BeautifulSoup def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: str = "AAPL" ): SCREAMING_SNAKE_CASE__ = f'''https://in.finance.yahoo.com/quote/{symbol}?s={symbol}''' SCREAMING_SNAKE_CASE__ = BeautifulSoup(requests.get(UpperCamelCase__ ).text , """html.parser""" ) SCREAMING_SNAKE_CASE__ = """My(6px) Pos(r) smartphone_Mt(6px)""" return soup.find("""div""" , class_=class_ ).find("""span""" ).text if __name__ == "__main__": for symbol in "AAPL AMZN IBM GOOG MSFT ORCL".split(): print(F'''Current {symbol:<4} stock price is {stock_price(symbol):>8}''')
6
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : Any , a_ : Union[str, Any]=13 , a_ : Any=7 , a_ : Any=True , a_ : Dict=True , a_ : Union[str, Any]=False , a_ : Tuple=True , a_ : str=99 , a_ : Tuple=64 , a_ : Tuple=5 , a_ : Union[str, Any]=4 , a_ : Dict=64 , a_ : Union[str, Any]="gelu" , a_ : Dict=0.1 , a_ : List[str]=0.1 , a_ : Dict=512 , a_ : Tuple=16 , a_ : str=2 , a_ : Any=0.02 , a_ : List[Any]=3 , a_ : Tuple=4 , a_ : Optional[int]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : int ): """simple docstring""" return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def A ( self : str ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def A ( self : Tuple , a_ : int , a_ : str , a_ : Optional[int] , a_ : List[Any] , a_ : str , a_ : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Any , a_ : int , a_ : Tuple , a_ : str , a_ : int , a_ : str , a_ : List[Any] ): """simple docstring""" __snake_case = MPNetForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Any , a_ : Any , a_ : int , a_ : Union[str, Any] , a_ : Dict , a_ : Optional[Any] , a_ : Any ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Optional[Any] , a_ : Any , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = MPNetForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict , a_ : List[str] , a_ : str , a_ : Union[str, Any] , a_ : str , a_ : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def A ( self : List[Any] ): """simple docstring""" __snake_case = MPNetModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel.from_pretrained("microsoft/mpnet-base" ) __snake_case = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a_ , atol=1e-4 ) )
69
0
"""simple docstring""" import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a = logging.get_logger(__name__) a = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } a = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } a = {'''facebook/blenderbot_small-90M''': 512} def _snake_case ( _snake_case : List[str] ) -> Union[str, Any]: '''simple docstring''' _A = set() _A = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _A = char _A = set(_snake_case ) return pairs class lowercase_ ( __lowerCAmelCase ): '''simple docstring''' UpperCAmelCase : List[str] = VOCAB_FILES_NAMES UpperCAmelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : List[Any] = ['''input_ids''', '''attention_mask'''] def __init__( self : int , _UpperCAmelCase : Tuple , _UpperCAmelCase : str , _UpperCAmelCase : Tuple="__start__" , _UpperCAmelCase : int="__end__" , _UpperCAmelCase : Optional[Any]="__unk__" , _UpperCAmelCase : List[str]="__null__" , **_UpperCAmelCase : str , ): super().__init__(unk_token=_UpperCAmelCase , bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , **_UpperCAmelCase ) with open(_UpperCAmelCase , encoding='utf-8' ) as vocab_handle: _A = json.load(_UpperCAmelCase ) _A = {v: k for k, v in self.encoder.items()} with open(_UpperCAmelCase , encoding='utf-8' ) as merges_handle: _A = merges_handle.read().split('\n' )[1:-1] _A = [tuple(merge.split() ) for merge in merges] _A = dict(zip(_UpperCAmelCase , range(len(_UpperCAmelCase ) ) ) ) _A = {} @property def lowerCAmelCase_ ( self : Optional[Any] ): return len(self.encoder ) def lowerCAmelCase_ ( self : Union[str, Any] ): return dict(self.encoder , **self.added_tokens_encoder ) def lowerCAmelCase_ ( self : Optional[int] , _UpperCAmelCase : str ): if token in self.cache: return self.cache[token] _A = re.sub('([.,!?()])' , r' \1' , _UpperCAmelCase ) _A = re.sub('(\')' , r' \1 ' , _UpperCAmelCase ) _A = re.sub(r'\s{2,}' , ' ' , _UpperCAmelCase ) if "\n" in token: _A = token.replace('\n' , ' __newln__' ) _A = token.split(' ' ) _A = [] for token in tokens: if not len(_UpperCAmelCase ): continue _A = token.lower() _A = tuple(_UpperCAmelCase ) _A = tuple(list(word[:-1] ) + [word[-1] + '</w>'] ) _A = get_pairs(_UpperCAmelCase ) if not pairs: words.append(_UpperCAmelCase ) continue while True: _A = min(_UpperCAmelCase , key=lambda _UpperCAmelCase : self.bpe_ranks.get(_UpperCAmelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break _A , _A = bigram _A = [] _A = 0 while i < len(_UpperCAmelCase ): try: _A = word.index(_UpperCAmelCase , _UpperCAmelCase ) new_word.extend(word[i:j] ) _A = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _A = tuple(_UpperCAmelCase ) _A = new_word if len(_UpperCAmelCase ) == 1: break else: _A = get_pairs(_UpperCAmelCase ) _A = '@@ '.join(_UpperCAmelCase ) _A = word[:-4] _A = word words.append(_UpperCAmelCase ) return " ".join(_UpperCAmelCase ) def lowerCAmelCase_ ( self : Optional[int] , _UpperCAmelCase : str ): _A = [] _A = re.findall(r'\S+\n?' , _UpperCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_UpperCAmelCase ).split(' ' ) ) ) return split_tokens def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : str ): _A = token.lower() return self.encoder.get(_UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : int ): return self.decoder.get(_UpperCAmelCase , self.unk_token ) def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : List[str] ): _A = ' '.join(_UpperCAmelCase ).replace('@@ ' , '' ).strip() return out_string def lowerCAmelCase_ ( self : Optional[int] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[str] = None ): if not os.path.isdir(_UpperCAmelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A = os.path.join( _UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _A = os.path.join( _UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(_UpperCAmelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_UpperCAmelCase , ensure_ascii=_UpperCAmelCase ) + '\n' ) _A = 0 with open(_UpperCAmelCase , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( F'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' ' Please check that the tokenizer is not corrupted!' ) _A = token_index writer.write(' '.join(_UpperCAmelCase ) + '\n' ) index += 1 return vocab_file, merge_file
7
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Optional[int]: return 1 / (1 + np.exp(-z )) def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict ) -> List[str]: return (-y * np.log(_UpperCAmelCase ) - (1 - y) * np.log(1 - h )).mean() def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any] ) -> Optional[Any]: __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) return np.sum(y * scores - np.log(1 + np.exp(_UpperCAmelCase ) ) ) def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str]=7_00_00 ) -> Union[str, Any]: __snake_case = np.zeros(x.shape[1] ) for iterations in range(_UpperCAmelCase ): __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = np.dot(x.T , h - y ) / y.size __snake_case = theta - alpha * gradient # updating the weights __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = cost_function(_UpperCAmelCase , _UpperCAmelCase ) if iterations % 1_00 == 0: print(F'''loss: {j} \t''' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": a : int = datasets.load_iris() a : int = iris.data[:, :2] a : Optional[Any] = (iris.target != 0) * 1 a : Tuple = 0.1 a : List[str] = logistic_reg(alpha, x, y, max_iterations=70_000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> Union[str, Any]: return sigmoid_function( np.dot(_UpperCAmelCase , _UpperCAmelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((a) , (a)) : Any = (x[:, 0].min(), x[:, 0].max()) ((a) , (a)) : Any = (x[:, 1].min(), x[:, 1].max()) ((a) , (a)) : Any = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) a : Optional[Any] = np.c_[xxa.ravel(), xxa.ravel()] a : List[Any] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
69
0
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): lowercase__ : Any = '''pt''' elif is_tf_available(): lowercase__ : Tuple = '''tf''' else: lowercase__ : Union[str, Any] = '''jax''' class SCREAMING_SNAKE_CASE (a__ , unittest.TestCase ): lowerCAmelCase = ByTaTokenizer lowerCAmelCase = False def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' super().setUp() __A : Any = ByTaTokenizer() tokenizer.save_pretrained(self.tmpdirname) @cached_property def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' return ByTaTokenizer.from_pretrained('google/byt5-small') def SCREAMING_SNAKE_CASE ( self , **_UpperCAmelCase): '''simple docstring''' return self.tokenizer_class.from_pretrained(self.tmpdirname , **_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase=False , _UpperCAmelCase=20 , _UpperCAmelCase=5): '''simple docstring''' __A : Optional[Any] = [] for i in range(len(_UpperCAmelCase)): try: __A : Optional[Any] = tokenizer.decode([i] , clean_up_tokenization_spaces=_UpperCAmelCase) except UnicodeDecodeError: pass toks.append((i, tok)) __A : List[str] = list(filter(lambda _UpperCAmelCase: re.match(R'^[ a-zA-Z]+$' , t[1]) , _UpperCAmelCase)) __A : Optional[Any] = list(filter(lambda _UpperCAmelCase: [t[0]] == tokenizer.encode(t[1] , add_special_tokens=_UpperCAmelCase) , _UpperCAmelCase)) if max_length is not None and len(_UpperCAmelCase) > max_length: __A : Optional[int] = toks[:max_length] if min_length is not None and len(_UpperCAmelCase) < min_length and len(_UpperCAmelCase) > 0: while len(_UpperCAmelCase) < min_length: __A : Tuple = toks + toks # toks_str = [t[1] for t in toks] __A : Tuple = [t[0] for t in toks] # Ensure consistency __A : Optional[Any] = tokenizer.decode(_UpperCAmelCase , clean_up_tokenization_spaces=_UpperCAmelCase) if " " not in output_txt and len(_UpperCAmelCase) > 1: __A : Union[str, Any] = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=_UpperCAmelCase) + ' ' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=_UpperCAmelCase) ) if with_prefix_space: __A : Dict = ' ' + output_txt __A : int = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase) return output_txt, output_ids def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[Any] = self.ta_base_tokenizer __A : Tuple = tokenizer(['hi</s>', 'I went to the gym</s>', '</s>']) __A : List[Any] = tokenizer(['hi', 'I went to the gym', '']) self.assertListEqual(batch_with_eos_added['input_ids'] , batch_without_eos_added['input_ids']) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[int] = self.ta_base_tokenizer __A : Any = 'Unicode €.' __A : Union[str, Any] = tokenizer(_UpperCAmelCase) __A : Optional[Any] = [88, 113, 108, 102, 114, 103, 104, 35, 229, 133, 175, 49, 1] self.assertEqual(encoded['input_ids'] , _UpperCAmelCase) # decoding __A : List[str] = tokenizer.decode(_UpperCAmelCase) self.assertEqual(_UpperCAmelCase , 'Unicode €.</s>') __A : Any = tokenizer('e è é ê ë') __A : List[str] = [104, 35, 198, 171, 35, 198, 172, 35, 198, 173, 35, 198, 174, 1] self.assertEqual(encoded['input_ids'] , _UpperCAmelCase) # decoding __A : Optional[int] = tokenizer.decode(_UpperCAmelCase) self.assertEqual(_UpperCAmelCase , 'e è é ê ë</s>') # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('e è é ê ë')) , 'e è é ê ë</s>') def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Union[str, Any] = self.ta_base_tokenizer __A : Optional[Any] = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] # fmt: off __A : Tuple = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 1, 0] # fmt: on __A : str = tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , return_tensors=_UpperCAmelCase) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase) if FRAMEWORK != "jax": __A : Optional[Any] = list(batch.input_ids.numpy()[0]) else: __A : str = list(batch.input_ids.tolist()[0]) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase) self.assertEqual((2, 37) , batch.input_ids.shape) self.assertEqual((2, 37) , batch.attention_mask.shape) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[Any] = self.ta_base_tokenizer __A : str = ['A long paragraph for summarization.', 'Another paragraph for summarization.'] __A : Tuple = tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , return_tensors=_UpperCAmelCase) # check if input_ids are returned and no decoder_input_ids self.assertIn('input_ids' , _UpperCAmelCase) self.assertIn('attention_mask' , _UpperCAmelCase) self.assertNotIn('decoder_input_ids' , _UpperCAmelCase) self.assertNotIn('decoder_attention_mask' , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Tuple = self.ta_base_tokenizer __A : Any = [ 'Summary of the text.', 'Another summary.', ] __A : Optional[Any] = tokenizer( text_target=_UpperCAmelCase , max_length=32 , padding='max_length' , truncation=_UpperCAmelCase , return_tensors=_UpperCAmelCase) self.assertEqual(32 , targets['input_ids'].shape[1]) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : str = self.ta_base_tokenizer __A : Optional[Any] = ['A long paragraph for summarization. </s>'] __A : Dict = ['Summary of the text. </s>'] # fmt: off __A : str = [68, 35, 111, 114, 113, 106, 35, 115, 100, 117, 100, 106, 117, 100, 115, 107, 35, 105, 114, 117, 35, 118, 120, 112, 112, 100, 117, 108, 125, 100, 119, 108, 114, 113, 49, 35, 1] __A : List[str] = [86, 120, 112, 112, 100, 117, 124, 35, 114, 105, 35, 119, 107, 104, 35, 119, 104, 123, 119, 49, 35, 1] # fmt: on __A : Any = tokenizer(_UpperCAmelCase , text_target=_UpperCAmelCase) self.assertEqual(_UpperCAmelCase , batch['input_ids'][0]) self.assertEqual(_UpperCAmelCase , batch['labels'][0]) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): self.assertNotEqual(tokenizer.model_max_length , 42) # Now let's start the test __A : List[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): # Isolate this from the other tests because we save additional tokens/etc __A : Dict = tempfile.mkdtemp() __A : Dict = ' He is very happy, UNwant\u00E9d,running' __A : List[str] = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase) tokenizer.save_pretrained(_UpperCAmelCase) __A : Optional[Any] = tokenizer.__class__.from_pretrained(_UpperCAmelCase) __A : Dict = after_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase) shutil.rmtree(_UpperCAmelCase) __A : List[Any] = self.get_tokenizers(model_max_length=42) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): # Isolate this from the other tests because we save additional tokens/etc __A : List[str] = tempfile.mkdtemp() __A : str = ' He is very happy, UNwant\u00E9d,running' tokenizer.add_tokens(['bim', 'bambam']) __A : List[str] = tokenizer.additional_special_tokens additional_special_tokens.append('new_additional_special_token') tokenizer.add_special_tokens({'additional_special_tokens': additional_special_tokens}) __A : Union[str, Any] = tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase) tokenizer.save_pretrained(_UpperCAmelCase) __A : Dict = tokenizer.__class__.from_pretrained(_UpperCAmelCase) __A : Optional[Any] = after_tokenizer.encode(_UpperCAmelCase , add_special_tokens=_UpperCAmelCase) self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase) self.assertIn('new_additional_special_token' , after_tokenizer.additional_special_tokens) self.assertEqual(after_tokenizer.model_max_length , 42) __A : str = tokenizer.__class__.from_pretrained(_UpperCAmelCase , model_max_length=43) self.assertEqual(tokenizer.model_max_length , 43) shutil.rmtree(_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : List[Any] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer())) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer())) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_UpperCAmelCase) with open(os.path.join(_UpperCAmelCase , 'special_tokens_map.json') , encoding='utf-8') as json_file: __A : Tuple = json.load(_UpperCAmelCase) with open(os.path.join(_UpperCAmelCase , 'tokenizer_config.json') , encoding='utf-8') as json_file: __A : List[Any] = json.load(_UpperCAmelCase) __A : str = [F'<extra_id_{i}>' for i in range(125)] __A : Union[str, Any] = added_tokens_extra_ids + [ 'an_additional_special_token' ] __A : List[Any] = added_tokens_extra_ids + [ 'an_additional_special_token' ] with open(os.path.join(_UpperCAmelCase , 'special_tokens_map.json') , 'w' , encoding='utf-8') as outfile: json.dump(_UpperCAmelCase , _UpperCAmelCase) with open(os.path.join(_UpperCAmelCase , 'tokenizer_config.json') , 'w' , encoding='utf-8') as outfile: json.dump(_UpperCAmelCase , _UpperCAmelCase) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files __A : Any = tokenizer_class.from_pretrained( _UpperCAmelCase , ) self.assertIn( 'an_additional_special_token' , tokenizer_without_change_in_init.additional_special_tokens) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( ['an_additional_special_token'] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['an_additional_special_token'])) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained __A : Union[str, Any] = added_tokens_extra_ids + [AddedToken('a_new_additional_special_token' , lstrip=_UpperCAmelCase)] __A : int = tokenizer_class.from_pretrained( _UpperCAmelCase , additional_special_tokens=_UpperCAmelCase , ) self.assertIn('a_new_additional_special_token' , tokenizer.additional_special_tokens) self.assertEqual( ['a_new_additional_special_token'] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['a_new_additional_special_token'])) , ) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[int] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer())) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer())) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(_UpperCAmelCase) __A : Optional[Any] = tokenizer_class.from_pretrained(_UpperCAmelCase) self.assertTrue(tokenizer.decode([255]) == '') def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' pass def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Dict = self.get_tokenizers(fast=_UpperCAmelCase , do_lower_case=_UpperCAmelCase) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): __A : List[str] = ['t', 'h', 'i', 's', ' ', 'i', 's', ' ', 'a', ' ', 't', 'e', 'x', 't', '</s>'] __A : Tuple = tokenizer.convert_tokens_to_string(_UpperCAmelCase) self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self): '''simple docstring''' __A : Optional[Any] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}'): __A : Tuple = [ 'bos_token', 'eos_token', 'unk_token', 'sep_token', 'pad_token', 'cls_token', 'mask_token', ] __A : Optional[int] = 0 __A : List[Any] = tokenizer.convert_ids_to_tokens( _UpperCAmelCase , skip_special_tokens=_UpperCAmelCase) for attr in attributes_list: setattr(_UpperCAmelCase , attr + '_id' , _UpperCAmelCase) self.assertEqual(getattr(_UpperCAmelCase , _UpperCAmelCase) , _UpperCAmelCase) self.assertEqual(getattr(_UpperCAmelCase , attr + '_id') , _UpperCAmelCase) setattr(_UpperCAmelCase , attr + '_id' , _UpperCAmelCase) self.assertEqual(getattr(_UpperCAmelCase , _UpperCAmelCase) , _UpperCAmelCase) self.assertEqual(getattr(_UpperCAmelCase , attr + '_id') , _UpperCAmelCase) setattr(_UpperCAmelCase , 'additional_special_tokens_ids' , []) self.assertListEqual(getattr(_UpperCAmelCase , 'additional_special_tokens') , []) self.assertListEqual(getattr(_UpperCAmelCase , 'additional_special_tokens_ids') , []) setattr(_UpperCAmelCase , 'additional_special_tokens_ids' , [token_id_to_test_setters]) self.assertListEqual(getattr(_UpperCAmelCase , 'additional_special_tokens') , [token_to_test_setters]) self.assertListEqual(getattr(_UpperCAmelCase , 'additional_special_tokens_ids') , [token_id_to_test_setters])
8
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
69
0
import cva import numpy as np class __lowerCAmelCase : """simple docstring""" def __init__( self : Union[str, Any] , _snake_case : float , _snake_case : int ): """simple docstring""" if k in (0.04, 0.06): A__ = k A__ = window_size else: raise ValueError('invalid k value' ) def __str__( self : Any ): """simple docstring""" return str(self.k ) def _a ( self : Union[str, Any] , _snake_case : str ): """simple docstring""" A__ = cva.imread(_snake_case , 0 ) A__ , A__ = img.shape A__ = [] A__ = img.copy() A__ = cva.cvtColor(_snake_case , cva.COLOR_GRAY2RGB ) A__ , A__ = np.gradient(_snake_case ) A__ = dx**2 A__ = dy**2 A__ = dx * dy A__ = 0.04 A__ = self.window_size // 2 for y in range(_snake_case , h - offset ): for x in range(_snake_case , w - offset ): A__ = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ = (wxx * wyy) - (wxy**2) A__ = wxx + wyy A__ = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 2_55 ) return color_img, corner_list if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = HarrisCorner(0.04, 3) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
9
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
0
def _snake_case ( __snake_case = 10**12 ): _UpperCamelCase = 1 _UpperCamelCase = 0 _UpperCamelCase = 1 _UpperCamelCase = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(f'{solution() = }')
10
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Tuple = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """autoformer""" __SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : List[Any] , a_ : Optional[int] = None , a_ : Optional[int] = None , a_ : str = "student_t" , a_ : str = "nll" , a_ : int = 1 , a_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , a_ : bool = True , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : Optional[List[int]] = None , a_ : Optional[List[int]] = None , a_ : int = 64 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 32 , a_ : int = 32 , a_ : str = "gelu" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : int = 100 , a_ : float = 0.02 , a_ : bool = True , a_ : Union[str, Any]=True , a_ : int = 10 , a_ : int = 25 , a_ : int = 3 , **a_ : Tuple , ): """simple docstring""" __snake_case = prediction_length __snake_case = context_length if context_length is not None else prediction_length __snake_case = distribution_output __snake_case = loss __snake_case = input_size __snake_case = num_time_features __snake_case = lags_sequence __snake_case = scaling __snake_case = num_dynamic_real_features __snake_case = num_static_real_features __snake_case = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case = cardinality else: __snake_case = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case = embedding_dimension else: __snake_case = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case = num_parallel_samples # Transformer architecture configuration __snake_case = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case = d_model __snake_case = encoder_attention_heads __snake_case = decoder_attention_heads __snake_case = encoder_ffn_dim __snake_case = decoder_ffn_dim __snake_case = encoder_layers __snake_case = decoder_layers __snake_case = dropout __snake_case = attention_dropout __snake_case = activation_dropout __snake_case = encoder_layerdrop __snake_case = decoder_layerdrop __snake_case = activation_function __snake_case = init_std __snake_case = use_cache # Autoformer __snake_case = label_length __snake_case = moving_average __snake_case = autocorrelation_factor super().__init__(is_encoder_decoder=a_ , **a_ ) @property def A ( self : Optional[int] ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
69
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor lowercase_ = logging.get_logger(__name__) class __A ( A ): '''simple docstring''' def __init__(self , *A , **A ) -> None: """simple docstring""" warnings.warn( '''The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use CLIPImageProcessor instead.''' , A , ) super().__init__(*A , **A )
11
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
0
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _snake_case : def __init__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=32 , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=10 , SCREAMING_SNAKE_CASE_=[10, 20, 30, 40] , SCREAMING_SNAKE_CASE_=[1, 1, 2, 1] , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_="relu" , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=None , ): '''simple docstring''' lowercase__ : Any = parent lowercase__ : Any = batch_size lowercase__ : Dict = image_size lowercase__ : Union[str, Any] = num_channels lowercase__ : Optional[Any] = embeddings_size lowercase__ : Optional[Any] = hidden_sizes lowercase__ : Any = depths lowercase__ : Optional[int] = is_training lowercase__ : Optional[int] = use_labels lowercase__ : Optional[int] = hidden_act lowercase__ : Dict = num_labels lowercase__ : str = scope lowercase__ : Optional[int] = len(SCREAMING_SNAKE_CASE_) def lowercase__ ( self): '''simple docstring''' lowercase__ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) lowercase__ : Union[str, Any] = None if self.use_labels: lowercase__ : Any = ids_tensor([self.batch_size] , self.num_labels) lowercase__ : str = self.get_config() return config, pixel_values, labels def lowercase__ ( self): '''simple docstring''' return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def lowercase__ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_): '''simple docstring''' lowercase__ : str = TFResNetModel(config=SCREAMING_SNAKE_CASE_) lowercase__ : int = model(SCREAMING_SNAKE_CASE_) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowercase__ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_): '''simple docstring''' lowercase__ : Union[str, Any] = self.num_labels lowercase__ : List[Any] = TFResNetForImageClassification(SCREAMING_SNAKE_CASE_) lowercase__ : int = model(SCREAMING_SNAKE_CASE_ , labels=SCREAMING_SNAKE_CASE_) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels)) def lowercase__ ( self): '''simple docstring''' lowercase__ : Tuple = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = config_and_inputs lowercase__ : List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _snake_case ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): __lowerCAmelCase : Any = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () __lowerCAmelCase : Optional[Any] = ( {'feature-extraction': TFResNetModel, 'image-classification': TFResNetForImageClassification} if is_tf_available() else {} ) __lowerCAmelCase : Optional[Any] = False __lowerCAmelCase : Optional[Any] = False __lowerCAmelCase : List[str] = False __lowerCAmelCase : str = False __lowerCAmelCase : List[Any] = False def lowercase__ ( self): '''simple docstring''' lowercase__ : Tuple = TFResNetModelTester(self) lowercase__ : str = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , has_text_modality=SCREAMING_SNAKE_CASE_) def lowercase__ ( self): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase__ ( self): '''simple docstring''' return @unittest.skip(reason="""ResNet does not use inputs_embeds""") def lowercase__ ( self): '''simple docstring''' pass @unittest.skip(reason="""ResNet does not support input and output embeddings""") def lowercase__ ( self): '''simple docstring''' pass def lowercase__ ( self): '''simple docstring''' lowercase__ , lowercase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ : Union[str, Any] = model_class(SCREAMING_SNAKE_CASE_) lowercase__ : Tuple = inspect.signature(model.call) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ : List[Any] = [*signature.parameters.keys()] lowercase__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE_) def lowercase__ ( self): '''simple docstring''' lowercase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_) def lowercase__ ( self): '''simple docstring''' def check_hidden_states_output(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_): lowercase__ : int = model_class(SCREAMING_SNAKE_CASE_) lowercase__ : Dict = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_)) lowercase__ : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowercase__ : Optional[Any] = self.model_tester.num_stages self.assertEqual(len(SCREAMING_SNAKE_CASE_) , expected_num_stages + 1) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowercase__ , lowercase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() lowercase__ : int = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: lowercase__ : Any = layer_type lowercase__ : str = True check_hidden_states_output(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowercase__ : Dict = True check_hidden_states_output(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_) def lowercase__ ( self): '''simple docstring''' lowercase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE_) @slow def lowercase__ ( self): '''simple docstring''' for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : List[str] = TFResNetModel.from_pretrained(SCREAMING_SNAKE_CASE_) self.assertIsNotNone(SCREAMING_SNAKE_CASE_) def UpperCamelCase ( ) -> Dict: '''simple docstring''' lowercase__ : str = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _snake_case ( unittest.TestCase ): @cached_property def lowercase__ ( self): '''simple docstring''' return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) if is_vision_available() else None ) @slow def lowercase__ ( self): '''simple docstring''' lowercase__ : Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0]) lowercase__ : Any = self.default_image_processor lowercase__ : Any = prepare_img() lowercase__ : List[str] = image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors="""tf""") # forward pass lowercase__ : List[Any] = model(**SCREAMING_SNAKE_CASE_) # verify the logits lowercase__ : List[str] = tf.TensorShape((1, 10_00)) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE_) lowercase__ : Optional[Any] = tf.constant([-1_1.1_0_6_9, -9.7_8_7_7, -8.3_7_7_7]) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , SCREAMING_SNAKE_CASE_ , atol=1E-4))
12
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 a : Tuple = get_tests_dir('''fixtures''') a : Dict = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : int = get_tests_dir('''fixtures/dummy-config.json''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Tuple ): """simple docstring""" __snake_case = 0 def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __snake_case = AutoFeatureExtractor.from_pretrained(a_ ).to_dict() config_dict.pop("feature_extractor_type" ) __snake_case = WavaVecaFeatureExtractor(**a_ ) # save in new folder model_config.save_pretrained(a_ ) config.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) # make sure private variable is not incorrectly saved __snake_case = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with self.assertRaisesRegex( a_ , "bert-base is not a local folder and is not a valid model identifier" ): __snake_case = AutoFeatureExtractor.from_pretrained("bert-base" ) def A ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( a_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __snake_case = AutoFeatureExtractor.from_pretrained(a_ , revision="aaaaaa" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( a_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): __snake_case = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ , trust_remote_code=a_ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def A ( self : int ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoFeatureExtractor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def A ( self : Dict ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # If remote code is not set, the default is to use local __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(a_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
69
0
'''simple docstring''' import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def UpperCAmelCase__ ( UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : List[Any]=10_24 , UpperCAmelCase_ : List[Any]=10_24 , UpperCAmelCase_ : List[str]=False , **UpperCAmelCase_ : str ) -> Dict: __lowerCamelCase : str = AutoTokenizer.from_pretrained(UpperCAmelCase_ ) __lowerCamelCase : Union[str, Any] = SeqaSeqDataset(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , type_path='train' , **UpperCAmelCase_ ) __lowerCamelCase : List[str] = tok.pad_token_id def get_lens(UpperCAmelCase_ : Optional[Any] ): __lowerCamelCase : Union[str, Any] = tqdm( DataLoader(UpperCAmelCase_ , batch_size=5_12 , num_workers=8 , shuffle=UpperCAmelCase_ , collate_fn=ds.collate_fn ) , desc=str(ds.len_file ) , ) __lowerCamelCase : Optional[Any] = [] for batch in dl: __lowerCamelCase : int = batch['input_ids'].ne(UpperCAmelCase_ ).sum(1 ).tolist() __lowerCamelCase : Any = batch['labels'].ne(UpperCAmelCase_ ).sum(1 ).tolist() if consider_target: for src, tgt in zip(UpperCAmelCase_ , UpperCAmelCase_ ): max_lens.append(max(UpperCAmelCase_ , UpperCAmelCase_ ) ) else: max_lens.extend(UpperCAmelCase_ ) return max_lens __lowerCamelCase : int = get_lens(UpperCAmelCase_ ) __lowerCamelCase : Any = SeqaSeqDataset(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , type_path='val' , **UpperCAmelCase_ ) __lowerCamelCase : Union[str, Any] = get_lens(UpperCAmelCase_ ) pickle_save(UpperCAmelCase_ , train_ds.len_file ) pickle_save(UpperCAmelCase_ , val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
13
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
0
def __UpperCAmelCase ( __a : int ,__a : int ) -> str: """simple docstring""" if not isinstance(__a ,__a ): raise ValueError('''iterations must be defined as integers''' ) if not isinstance(__a ,__a ) or not number >= 1: raise ValueError( '''starting number must be and integer and be more than 0''' ) if not iterations >= 1: raise ValueError('''Iterations must be done more than 0 times to play FizzBuzz''' ) _a : List[Any] = '''''' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(__a ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
14
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
0
import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class A ( unittest.TestCase ): '''simple docstring''' def __init__(self : List[str] , _UpperCAmelCase : Dict , _UpperCAmelCase : List[Any]=13 , _UpperCAmelCase : List[str]=7 , _UpperCAmelCase : Union[str, Any]=True , _UpperCAmelCase : str=True , _UpperCAmelCase : Any=True , _UpperCAmelCase : Dict=True , _UpperCAmelCase : List[Any]=99 , _UpperCAmelCase : Tuple=32 , _UpperCAmelCase : Dict=5 , _UpperCAmelCase : Optional[int]=4 , _UpperCAmelCase : str=37 , _UpperCAmelCase : str="gelu" , _UpperCAmelCase : int=0.1 , _UpperCAmelCase : int=0.1 , _UpperCAmelCase : Dict=512 , _UpperCAmelCase : List[str]=16 , _UpperCAmelCase : Any=2 , _UpperCAmelCase : Optional[int]=0.02 , _UpperCAmelCase : Optional[Any]=4 , ) -> Any: """simple docstring""" lowercase__ = parent lowercase__ = batch_size lowercase__ = seq_length lowercase__ = is_training lowercase__ = use_attention_mask lowercase__ = use_token_type_ids lowercase__ = use_labels lowercase__ = vocab_size lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = max_position_embeddings lowercase__ = type_vocab_size lowercase__ = type_sequence_label_size lowercase__ = initializer_range lowercase__ = num_choices def lowerCamelCase__ (self : Dict ) -> List[str]: """simple docstring""" lowercase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ = None if self.use_attention_mask: lowercase__ = random_attention_mask([self.batch_size, self.seq_length] ) lowercase__ = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=_UpperCAmelCase , ) return config, input_ids, attention_mask def lowerCamelCase__ (self : Any ) -> List[Any]: """simple docstring""" lowercase__ = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ = config_and_inputs lowercase__ = {"""input_ids""": input_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class A ( UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' A__ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCamelCase__ (self : List[Any] ) -> Tuple: """simple docstring""" lowercase__ = FlaxDistilBertModelTester(self ) @slow def lowerCamelCase__ (self : List[str] ) -> List[Any]: """simple docstring""" for model_class_name in self.all_model_classes: lowercase__ = model_class_name.from_pretrained("""distilbert-base-uncased""" ) lowercase__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(_UpperCAmelCase ) @require_flax class A ( unittest.TestCase ): '''simple docstring''' @slow def lowerCamelCase__ (self : Optional[int] ) -> Optional[Any]: """simple docstring""" lowercase__ = FlaxDistilBertModel.from_pretrained("""distilbert-base-uncased""" ) lowercase__ = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) lowercase__ = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) lowercase__ = model(_UpperCAmelCase , attention_mask=_UpperCAmelCase )[0] lowercase__ = (1, 11, 768) self.assertEqual(output.shape , _UpperCAmelCase ) lowercase__ = np.array([[[-0.1_639, 0.3_299, 0.1_648], [-0.1_746, 0.3_289, 0.1_710], [-0.1_884, 0.3_357, 0.1_810]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , _UpperCAmelCase , atol=1E-4 ) )
15
'''simple docstring''' a : Dict = range(2, 20 + 1) a : Optional[int] = [10**k for k in range(ks[-1] + 1)] a : dict[int, dict[int, list[list[int]]]] = {} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] ) -> int: __snake_case = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) __snake_case = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) __snake_case , __snake_case = 0, 0 __snake_case = n - i __snake_case = memo.get(_UpperCAmelCase ) if sub_memo is not None: __snake_case = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over __snake_case = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __snake_case = _k break if max_jump >= 0: __snake_case , __snake_case , __snake_case = jumps[max_jump] # since the difference between jumps is cached, add c __snake_case = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: __snake_case = [] else: __snake_case = {c: []} __snake_case = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __snake_case , __snake_case = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __snake_case , __snake_case = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped __snake_case = sub_memo[c] # keep jumps sorted by # of terms skipped __snake_case = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __snake_case = i __snake_case , __snake_case , __snake_case = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __snake_case = ds_c + ds_b diff += addend __snake_case = 0 for j in range(_UpperCAmelCase ): __snake_case = a_i[j] + addend __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Tuple: for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case = digits[j] + addend if s >= 10: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) __snake_case = addend // 10 + quotient else: __snake_case = s __snake_case = addend // 10 if addend == 0: break while addend > 0: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int = 10**15 ) -> int: __snake_case = [1] __snake_case = 1 __snake_case = 0 while True: __snake_case , __snake_case = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break __snake_case = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
69
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() __A : Optional[Any] = logging.get_logger(__name__) def __a ( A__ : int ): SCREAMING_SNAKE_CASE = DPTConfig() if "large" in checkpoint_url: SCREAMING_SNAKE_CASE = 1024 SCREAMING_SNAKE_CASE = 4096 SCREAMING_SNAKE_CASE = 24 SCREAMING_SNAKE_CASE = 16 SCREAMING_SNAKE_CASE = [5, 11, 17, 23] SCREAMING_SNAKE_CASE = [256, 512, 1024, 1024] SCREAMING_SNAKE_CASE = (1, 384, 384) if "ade" in checkpoint_url: SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = 150 SCREAMING_SNAKE_CASE = "huggingface/label-files" SCREAMING_SNAKE_CASE = "ade20k-id2label.json" SCREAMING_SNAKE_CASE = json.load(open(cached_download(hf_hub_url(A__ , A__ , repo_type="dataset" ) ) , "r" ) ) SCREAMING_SNAKE_CASE = {int(A__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE = idalabel SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} SCREAMING_SNAKE_CASE = [1, 150, 480, 480] return config, expected_shape def __a ( A__ : Optional[Any] ): SCREAMING_SNAKE_CASE = ["pretrained.model.head.weight", "pretrained.model.head.bias"] for k in ignore_keys: state_dict.pop(A__ , A__ ) def __a ( A__ : Tuple ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): SCREAMING_SNAKE_CASE = name.replace("pretrained.model" , "dpt.encoder" ) if "pretrained.model" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.model" , "dpt.embeddings" ) if "patch_embed" in name: SCREAMING_SNAKE_CASE = name.replace("patch_embed" , "patch_embeddings" ) if "pos_embed" in name: SCREAMING_SNAKE_CASE = name.replace("pos_embed" , "position_embeddings" ) if "attn.proj" in name: SCREAMING_SNAKE_CASE = name.replace("attn.proj" , "attention.output.dense" ) if "proj" in name and "project" not in name: SCREAMING_SNAKE_CASE = name.replace("proj" , "projection" ) if "blocks" in name: SCREAMING_SNAKE_CASE = name.replace("blocks" , "layer" ) if "mlp.fc1" in name: SCREAMING_SNAKE_CASE = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: SCREAMING_SNAKE_CASE = name.replace("mlp.fc2" , "output.dense" ) if "norm1" in name: SCREAMING_SNAKE_CASE = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: SCREAMING_SNAKE_CASE = name.replace("norm2" , "layernorm_after" ) if "scratch.output_conv" in name: SCREAMING_SNAKE_CASE = name.replace("scratch.output_conv" , "head" ) if "scratch" in name: SCREAMING_SNAKE_CASE = name.replace("scratch" , "neck" ) if "layer1_rn" in name: SCREAMING_SNAKE_CASE = name.replace("layer1_rn" , "convs.0" ) if "layer2_rn" in name: SCREAMING_SNAKE_CASE = name.replace("layer2_rn" , "convs.1" ) if "layer3_rn" in name: SCREAMING_SNAKE_CASE = name.replace("layer3_rn" , "convs.2" ) if "layer4_rn" in name: SCREAMING_SNAKE_CASE = name.replace("layer4_rn" , "convs.3" ) if "refinenet" in name: SCREAMING_SNAKE_CASE = int(name[len("neck.refinenet" ) : len("neck.refinenet" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 SCREAMING_SNAKE_CASE = name.replace(F"refinenet{layer_idx}" , F"fusion_stage.layers.{abs(layer_idx-4 )}" ) if "out_conv" in name: SCREAMING_SNAKE_CASE = name.replace("out_conv" , "projection" ) if "resConfUnit1" in name: SCREAMING_SNAKE_CASE = name.replace("resConfUnit1" , "residual_layer1" ) if "resConfUnit2" in name: SCREAMING_SNAKE_CASE = name.replace("resConfUnit2" , "residual_layer2" ) if "conv1" in name: SCREAMING_SNAKE_CASE = name.replace("conv1" , "convolution1" ) if "conv2" in name: SCREAMING_SNAKE_CASE = name.replace("conv2" , "convolution2" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess1.0.project.0" , "neck.reassemble_stage.readout_projects.0.0" ) if "pretrained.act_postprocess2.0.project.0" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess2.0.project.0" , "neck.reassemble_stage.readout_projects.1.0" ) if "pretrained.act_postprocess3.0.project.0" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess3.0.project.0" , "neck.reassemble_stage.readout_projects.2.0" ) if "pretrained.act_postprocess4.0.project.0" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess4.0.project.0" , "neck.reassemble_stage.readout_projects.3.0" ) # resize blocks if "pretrained.act_postprocess1.3" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess1.3" , "neck.reassemble_stage.layers.0.projection" ) if "pretrained.act_postprocess1.4" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess1.4" , "neck.reassemble_stage.layers.0.resize" ) if "pretrained.act_postprocess2.3" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess2.3" , "neck.reassemble_stage.layers.1.projection" ) if "pretrained.act_postprocess2.4" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess2.4" , "neck.reassemble_stage.layers.1.resize" ) if "pretrained.act_postprocess3.3" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess3.3" , "neck.reassemble_stage.layers.2.projection" ) if "pretrained.act_postprocess4.3" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess4.3" , "neck.reassemble_stage.layers.3.projection" ) if "pretrained.act_postprocess4.4" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained.act_postprocess4.4" , "neck.reassemble_stage.layers.3.resize" ) if "pretrained" in name: SCREAMING_SNAKE_CASE = name.replace("pretrained" , "dpt" ) if "bn" in name: SCREAMING_SNAKE_CASE = name.replace("bn" , "batch_norm" ) if "head" in name: SCREAMING_SNAKE_CASE = name.replace("head" , "head.head" ) if "encoder.norm" in name: SCREAMING_SNAKE_CASE = name.replace("encoder.norm" , "layernorm" ) if "auxlayer" in name: SCREAMING_SNAKE_CASE = name.replace("auxlayer" , "auxiliary_head.head" ) return name def __a ( A__ : Dict , A__ : List[Any] ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) SCREAMING_SNAKE_CASE = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.weight" ) SCREAMING_SNAKE_CASE = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE = in_proj_weight[: config.hidden_size, :] SCREAMING_SNAKE_CASE = in_proj_bias[: config.hidden_size] SCREAMING_SNAKE_CASE = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] SCREAMING_SNAKE_CASE = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] SCREAMING_SNAKE_CASE = in_proj_weight[ -config.hidden_size :, : ] SCREAMING_SNAKE_CASE = in_proj_bias[-config.hidden_size :] def __a ( ): SCREAMING_SNAKE_CASE = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def __a ( A__ : Tuple , A__ : Tuple , A__ : int , A__ : Tuple ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = get_dpt_config(A__ ) # load original state_dict from URL SCREAMING_SNAKE_CASE = torch.hub.load_state_dict_from_url(A__ , map_location="cpu" ) # remove certain keys remove_ignore_keys_(A__ ) # rename keys for key in state_dict.copy().keys(): SCREAMING_SNAKE_CASE = state_dict.pop(A__ ) SCREAMING_SNAKE_CASE = val # read in qkv matrices read_in_q_k_v(A__ , A__ ) # load HuggingFace model SCREAMING_SNAKE_CASE = DPTForSemanticSegmentation(A__ ) if "ade" in checkpoint_url else DPTForDepthEstimation(A__ ) model.load_state_dict(A__ ) model.eval() # Check outputs on an image SCREAMING_SNAKE_CASE = 480 if "ade" in checkpoint_url else 384 SCREAMING_SNAKE_CASE = DPTImageProcessor(size=A__ ) SCREAMING_SNAKE_CASE = prepare_img() SCREAMING_SNAKE_CASE = image_processor(A__ , return_tensors="pt" ) # forward pass SCREAMING_SNAKE_CASE = model(**A__ ).logits if "ade" in checkpoint_url else model(**A__ ).predicted_depth # Assert logits SCREAMING_SNAKE_CASE = torch.tensor([[6.3_1_9_9, 6.3_6_2_9, 6.4_1_4_8], [6.3_8_5_0, 6.3_6_1_5, 6.4_1_6_6], [6.3_5_1_9, 6.3_1_7_6, 6.3_5_7_5]] ) if "ade" in checkpoint_url: SCREAMING_SNAKE_CASE = torch.tensor([[4.0_4_8_0, 4.2_4_2_0, 4.4_3_6_0], [4.3_1_2_4, 4.5_6_9_3, 4.8_2_6_1], [4.5_7_6_8, 4.8_9_6_5, 5.2_1_6_3]] ) assert outputs.shape == torch.Size(A__ ) assert ( torch.allclose(outputs[0, 0, :3, :3] , A__ , atol=1E-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3] , A__ ) ) Path(A__ ).mkdir(exist_ok=A__ ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(A__ ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(A__ ) if push_to_hub: print("Pushing model to hub..." ) model.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="nielsr" , commit_message="Add model" , use_temp_dir=A__ , ) image_processor.push_to_hub( repo_path_or_name=Path(A__ , A__ ) , organization="nielsr" , commit_message="Add image processor" , use_temp_dir=A__ , ) if __name__ == "__main__": __A : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) __A : Optional[int] = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
16
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
0
import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class lowerCamelCase_ ( _lowercase , unittest.TestCase ): _lowercase : Optional[int] = BioGptTokenizer _lowercase : List[Any] = False def lowerCAmelCase_ ( self : str ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __A : Dict = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] __A : Tuple = dict(zip(__A , range(len(__A ) ) ) ) __A : List[str] = ["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] __A : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) __A : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def lowerCAmelCase_ ( self : Optional[Any] , __A : List[Any] ): __A : Optional[Any] = """lower newer""" __A : Optional[Any] = """lower newer""" return input_text, output_text def lowerCAmelCase_ ( self : Dict ): __A : int = BioGptTokenizer(self.vocab_file , self.merges_file ) __A : Any = """lower""" __A : str = ["""low""", """er</w>"""] __A : Optional[Any] = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) __A : Any = tokens + ["""<unk>"""] __A : Optional[int] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) @slow def lowerCAmelCase_ ( self : Tuple ): __A : List[str] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) __A : str = tokenizer.encode("""sequence builders""" , add_special_tokens=__A ) __A : Any = tokenizer.encode("""multi-sequence build""" , add_special_tokens=__A ) __A : Optional[int] = tokenizer.build_inputs_with_special_tokens(__A ) __A : Dict = tokenizer.build_inputs_with_special_tokens(__A , __A ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
17
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _SCREAMING_SNAKE_CASE = { "configuration_biogpt": ["BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BioGptConfig"], "tokenization_biogpt": ["BioGptTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ "BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST", "BioGptForCausalLM", "BioGptForTokenClassification", "BioGptForSequenceClassification", "BioGptModel", "BioGptPreTrainedModel", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
18
'''simple docstring''' import mpmath # for roots of unity import numpy as np class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , a_ : Optional[int]=None , a_ : int=None ): """simple docstring""" __snake_case = list(poly_a or [0] )[:] __snake_case = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() __snake_case = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() __snake_case = len(self.polyB ) # Add 0 to make lengths equal a power of 2 __snake_case = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform __snake_case = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product __snake_case = self.__multiply() def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" __snake_case = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB] # Corner case if len(a_ ) <= 1: return dft[0] # __snake_case = self.c_max_length // 2 while next_ncol > 0: __snake_case = [[] for i in range(a_ )] __snake_case = self.root**next_ncol # First half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update __snake_case = new_dft __snake_case = next_ncol // 2 return dft[0] def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.__dft("A" ) __snake_case = self.__dft("B" ) __snake_case = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT __snake_case = 2 while next_ncol <= self.c_max_length: __snake_case = [[] for i in range(a_ )] __snake_case = self.root ** (next_ncol // 2) __snake_case = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update __snake_case = new_inverse_c next_ncol *= 2 # Unpack __snake_case = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Optional[int] ): """simple docstring""" __snake_case = "A = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) __snake_case = "B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) __snake_case = "A*B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
69
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 42 lowercase__ = 42 def __init__( self , __a , __a) -> str: '''simple docstring''' super().__init__() self.register_modules(unet=__a , scheduler=__a) @torch.no_grad() def __call__( self , __a = 1 , __a = 20_00 , __a = None , __a = "pil" , __a = True , **__a , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' _UpperCamelCase = self.unet.config.sample_size _UpperCamelCase = (batch_size, 3, img_size, img_size) _UpperCamelCase = self.unet _UpperCamelCase = randn_tensor(__a , generator=__a) * self.scheduler.init_noise_sigma _UpperCamelCase = sample.to(self.device) self.scheduler.set_timesteps(__a) self.scheduler.set_sigmas(__a) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps)): _UpperCamelCase = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device) # correction step for _ in range(self.scheduler.config.correct_steps): _UpperCamelCase = self.unet(__a , __a).sample _UpperCamelCase = self.scheduler.step_correct(__a , __a , generator=__a).prev_sample # prediction step _UpperCamelCase = model(__a , __a).sample _UpperCamelCase = self.scheduler.step_pred(__a , __a , __a , generator=__a) _UpperCamelCase , _UpperCamelCase = output.prev_sample, output.prev_sample_mean _UpperCamelCase = sample_mean.clamp(0 , 1) _UpperCamelCase = sample.cpu().permute(0 , 2 , 3 , 1).numpy() if output_type == "pil": _UpperCamelCase = self.numpy_to_pil(__a) if not return_dict: return (sample,) return ImagePipelineOutput(images=__a)
19
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
0