code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def snake_case_ ( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Union[str, Any] ): """simple docstring""" lowercase_ : int = FunnelConfig.from_json_file(_lowerCAmelCase ) print(F'''Building PyTorch model from configuration: {config}''' ) lowercase_ : List[Any] = FunnelBaseModel(_lowerCAmelCase ) if base_model else FunnelModel(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _lowercase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) _lowercase : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
93
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase : Callable[[int | float], int | float] , _lowerCAmelCase : int | float , _lowerCAmelCase : int | float , _lowerCAmelCase : int = 100 , ): '''simple docstring''' lowercase__ : Dict = x_start lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length lowercase__ : Union[str, Any] = (x_end - x_start) / steps + xa lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowercase__ : Union[str, Any] = xa lowercase__ : int = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' return math.sin(10 * x ) print("f(x) = sin(10 * x)") print("The length of the curve from x = -10 to x = 10 is:") _UpperCamelCase : str = 10 while i <= 10_00_00: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
77
0
"""simple docstring""" import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class _SCREAMING_SNAKE_CASE ( A__ , unittest.TestCase ): UpperCAmelCase_ :List[str] = BioGptTokenizer UpperCAmelCase_ :int = False def __lowerCAmelCase ( self ) -> str: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase_ :str = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] lowerCAmelCase_ :Any = dict(zip(__A , range(len(__A ) ) ) ) lowerCAmelCase_ :str = ["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] lowerCAmelCase_ :Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) lowerCAmelCase_ :List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(__A ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(__A ) ) def __lowerCAmelCase ( self , __A ) -> Union[str, Any]: lowerCAmelCase_ :int = """lower newer""" lowerCAmelCase_ :str = """lower newer""" return input_text, output_text def __lowerCAmelCase ( self ) -> Optional[int]: lowerCAmelCase_ :Tuple = BioGptTokenizer(self.vocab_file , self.merges_file ) lowerCAmelCase_ :Dict = """lower""" lowerCAmelCase_ :List[str] = ["""low""", """er</w>"""] lowerCAmelCase_ :Any = tokenizer.tokenize(__A ) self.assertListEqual(__A , __A ) lowerCAmelCase_ :int = tokens + ["""<unk>"""] lowerCAmelCase_ :Any = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , __A ) @slow def __lowerCAmelCase ( self ) -> Dict: lowerCAmelCase_ :Optional[int] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) lowerCAmelCase_ :Dict = tokenizer.encode("""sequence builders""" , add_special_tokens=__A ) lowerCAmelCase_ :Any = tokenizer.encode("""multi-sequence build""" , add_special_tokens=__A ) lowerCAmelCase_ :Dict = tokenizer.build_inputs_with_special_tokens(__A ) lowerCAmelCase_ :Tuple = tokenizer.build_inputs_with_special_tokens(__A , __A ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
1
"""simple docstring""" import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 __UpperCAmelCase = get_tests_dir('fixtures/dummy-config.json') class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Dict: lowerCAmelCase_ :int = 0 def __lowerCAmelCase ( self ) -> List[str]: self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec("""transformers.models.auto""" ) ) def __lowerCAmelCase ( self ) -> Tuple: lowerCAmelCase_ :Tuple = AutoConfig.from_pretrained("""bert-base-uncased""" ) self.assertIsInstance(__A , __A ) def __lowerCAmelCase ( self ) -> Union[str, Any]: lowerCAmelCase_ :int = AutoConfig.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def __lowerCAmelCase ( self ) -> Any: lowerCAmelCase_ :Any = AutoConfig.from_pretrained(__A ) self.assertIsInstance(__A , __A ) def __lowerCAmelCase ( self ) -> Dict: lowerCAmelCase_ :int = AutoConfig.for_model("""roberta""" ) self.assertIsInstance(__A , __A ) def __lowerCAmelCase ( self ) -> Tuple: with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. lowerCAmelCase_ :int = os.path.join(__A , """fake-roberta""" ) os.makedirs(__A , exist_ok=__A ) with open(os.path.join(__A , """config.json""" ) , """w""" ) as f: f.write(json.dumps({} ) ) lowerCAmelCase_ :Any = AutoConfig.from_pretrained(__A ) self.assertEqual(type(__A ) , __A ) def __lowerCAmelCase ( self ) -> Optional[int]: try: AutoConfig.register("""custom""" , __A ) # Wrong model type will raise an error with self.assertRaises(__A ): AutoConfig.register("""model""" , __A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__A ): AutoConfig.register("""bert""" , __A ) # Now that the config is registered, it can be used as any other config with the auto-API lowerCAmelCase_ :Union[str, Any] = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__A ) lowerCAmelCase_ :Optional[int] = AutoConfig.from_pretrained(__A ) self.assertIsInstance(__A , __A ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def __lowerCAmelCase ( self ) -> Tuple: with self.assertRaisesRegex( __A , """bert-base is not a local folder and is not a valid model identifier""" ): lowerCAmelCase_ :List[str] = AutoConfig.from_pretrained("""bert-base""" ) def __lowerCAmelCase ( self ) -> Any: with self.assertRaisesRegex( __A , r"""aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)""" ): lowerCAmelCase_ :Dict = AutoConfig.from_pretrained(__A , revision="""aaaaaa""" ) def __lowerCAmelCase ( self ) -> int: with self.assertRaisesRegex( __A , """hf-internal-testing/no-config-test-repo does not appear to have a file named config.json.""" , ): lowerCAmelCase_ :Union[str, Any] = AutoConfig.from_pretrained("""hf-internal-testing/no-config-test-repo""" ) def __lowerCAmelCase ( self ) -> Tuple: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(__A ): lowerCAmelCase_ :Tuple = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" ) # If remote code is disabled, we can't load this config. with self.assertRaises(__A ): lowerCAmelCase_ :List[str] = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=__A ) lowerCAmelCase_ :str = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=__A ) self.assertEqual(config.__class__.__name__ , """NewModelConfig""" ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__A ) lowerCAmelCase_ :Dict = AutoConfig.from_pretrained(__A , trust_remote_code=__A ) self.assertEqual(reloaded_config.__class__.__name__ , """NewModelConfig""" ) def __lowerCAmelCase ( self ) -> int: class _SCREAMING_SNAKE_CASE ( A__ ): UpperCAmelCase_ :int = "new-model" try: AutoConfig.register("""new-model""" , __A ) # If remote code is not set, the default is to use local lowerCAmelCase_ :Any = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" ) self.assertEqual(config.__class__.__name__ , """NewModelConfigLocal""" ) # If remote code is disabled, we load the local one. lowerCAmelCase_ :Union[str, Any] = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=__A ) self.assertEqual(config.__class__.__name__ , """NewModelConfigLocal""" ) # If remote is enabled, we load from the Hub lowerCAmelCase_ :Optional[Any] = AutoConfig.from_pretrained("""hf-internal-testing/test_dynamic_model""" , trust_remote_code=__A ) self.assertEqual(config.__class__.__name__ , """NewModelConfig""" ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
1
1
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def __magic_name__ ( __UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' snake_case_ = VideoMAEConfig() set_architecture_configs(__UpperCAmelCase, __UpperCAmelCase ) if "finetuned" not in model_name: snake_case_ = False if "finetuned" in model_name: snake_case_ = '''huggingface/label-files''' if "kinetics" in model_name: snake_case_ = 400 snake_case_ = '''kinetics400-id2label.json''' elif "ssv2" in model_name: snake_case_ = 174 snake_case_ = '''something-something-v2-id2label.json''' else: raise ValueError('''Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.''' ) snake_case_ = json.load(open(hf_hub_download(__UpperCAmelCase, __UpperCAmelCase, repo_type='''dataset''' ), '''r''' ) ) snake_case_ = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} return config def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> List[str]: '''simple docstring''' if "small" in model_name: snake_case_ = 384 snake_case_ = 1536 snake_case_ = 12 snake_case_ = 16 snake_case_ = 12 snake_case_ = 3 snake_case_ = 192 snake_case_ = 768 elif "large" in model_name: snake_case_ = 1024 snake_case_ = 4096 snake_case_ = 24 snake_case_ = 16 snake_case_ = 12 snake_case_ = 8 snake_case_ = 512 snake_case_ = 2048 elif "huge" in model_name: snake_case_ = 1280 snake_case_ = 5120 snake_case_ = 32 snake_case_ = 16 snake_case_ = 12 snake_case_ = 8 snake_case_ = 640 snake_case_ = 2560 elif "base" not in model_name: raise ValueError('''Model name should include either "small", "base", "large", or "huge"''' ) def __magic_name__ ( __UpperCAmelCase ) -> List[Any]: '''simple docstring''' if "encoder." in name: snake_case_ = name.replace('''encoder.''', '''''' ) if "cls_token" in name: snake_case_ = name.replace('''cls_token''', '''videomae.embeddings.cls_token''' ) if "decoder_pos_embed" in name: snake_case_ = name.replace('''decoder_pos_embed''', '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case_ = name.replace('''pos_embed''', '''videomae.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case_ = name.replace('''patch_embed.proj''', '''videomae.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case_ = name.replace('''patch_embed.norm''', '''videomae.embeddings.norm''' ) if "decoder.blocks" in name: snake_case_ = name.replace('''decoder.blocks''', '''decoder.decoder_layers''' ) if "blocks" in name: snake_case_ = name.replace('''blocks''', '''videomae.encoder.layer''' ) if "attn.proj" in name: snake_case_ = name.replace('''attn.proj''', '''attention.output.dense''' ) if "attn" in name and "bias" not in name: snake_case_ = name.replace('''attn''', '''attention.self''' ) if "attn" in name: snake_case_ = name.replace('''attn''', '''attention.attention''' ) if "norm1" in name: snake_case_ = name.replace('''norm1''', '''layernorm_before''' ) if "norm2" in name: snake_case_ = name.replace('''norm2''', '''layernorm_after''' ) if "mlp.fc1" in name: snake_case_ = name.replace('''mlp.fc1''', '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case_ = name.replace('''mlp.fc2''', '''output.dense''' ) if "decoder_embed" in name: snake_case_ = name.replace('''decoder_embed''', '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case_ = name.replace('''decoder_norm''', '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case_ = name.replace('''decoder_pred''', '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: snake_case_ = name.replace('''norm.weight''', '''videomae.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: snake_case_ = name.replace('''norm.bias''', '''videomae.layernorm.bias''' ) if "head" in name and "decoder" not in name: snake_case_ = name.replace('''head''', '''classifier''' ) return name def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' for key in orig_state_dict.copy().keys(): snake_case_ = orig_state_dict.pop(__UpperCAmelCase ) if key.startswith('''encoder.''' ): snake_case_ = key.replace('''encoder.''', '''''' ) if "qkv" in key: snake_case_ = key.split('''.''' ) if key.startswith('''decoder.blocks''' ): snake_case_ = config.decoder_hidden_size snake_case_ = int(key_split[2] ) snake_case_ = '''decoder.decoder_layers.''' if "weight" in key: snake_case_ = val[:dim, :] snake_case_ = val[dim : dim * 2, :] snake_case_ = val[-dim:, :] else: snake_case_ = config.hidden_size snake_case_ = int(key_split[1] ) snake_case_ = '''videomae.encoder.layer.''' if "weight" in key: snake_case_ = val[:dim, :] snake_case_ = val[dim : dim * 2, :] snake_case_ = val[-dim:, :] else: snake_case_ = val return orig_state_dict def __magic_name__ ( ) -> List[Any]: '''simple docstring''' snake_case_ = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''', filename='''eating_spaghetti.npy''', repo_type='''dataset''' ) snake_case_ = np.load(__UpperCAmelCase ) return list(__UpperCAmelCase ) def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase ) -> Any: '''simple docstring''' snake_case_ = get_videomae_config(__UpperCAmelCase ) if "finetuned" in model_name: snake_case_ = VideoMAEForVideoClassification(__UpperCAmelCase ) else: snake_case_ = VideoMAEForPreTraining(__UpperCAmelCase ) # download original checkpoint, hosted on Google Drive snake_case_ = '''pytorch_model.bin''' gdown.cached_download(__UpperCAmelCase, __UpperCAmelCase, quiet=__UpperCAmelCase ) snake_case_ = torch.load(__UpperCAmelCase, map_location='''cpu''' ) if "model" in files: snake_case_ = files['''model'''] else: snake_case_ = files['''module'''] snake_case_ = convert_state_dict(__UpperCAmelCase, __UpperCAmelCase ) model.load_state_dict(__UpperCAmelCase ) model.eval() # verify model on basic input snake_case_ = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5], image_std=[0.5, 0.5, 0.5] ) snake_case_ = prepare_video() snake_case_ = image_processor(__UpperCAmelCase, return_tensors='''pt''' ) if "finetuned" not in model_name: snake_case_ = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''', filename='''bool_masked_pos.pt''' ) snake_case_ = torch.load(__UpperCAmelCase ) snake_case_ = model(**__UpperCAmelCase ) snake_case_ = outputs.logits snake_case_ = [ '''videomae-small-finetuned-kinetics''', '''videomae-small-finetuned-ssv2''', # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) '''videomae-base-short''', '''videomae-base-short-finetuned-kinetics''', '''videomae-base''', '''videomae-base-finetuned-kinetics''', '''videomae-large''', '''videomae-large-finetuned-kinetics''', '''videomae-huge-finetuned-kinetics''', # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) '''videomae-base-short-ssv2''', '''videomae-base-short-finetuned-ssv2''', '''videomae-base-ssv2''', '''videomae-base-finetuned-ssv2''', ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([-0.9_2_9_1, -0.4_0_6_1, -0.9_3_0_7] ) elif model_name == "videomae-small-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([0.2_6_7_1, -0.4_6_8_9, -0.8_2_3_5] ) elif model_name == "videomae-base": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_7_3_9, 0.7_9_6_8, 0.7_0_8_9], [0.6_7_0_1, 0.7_4_8_7, 0.6_2_0_9], [0.4_2_8_7, 0.5_1_5_8, 0.4_7_7_3]] ) elif model_name == "videomae-base-short": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_9_9_4, 0.9_6_1_2, 0.8_5_0_8], [0.7_4_0_1, 0.8_9_5_8, 0.8_3_0_2], [0.5_8_6_2, 0.7_4_6_8, 0.7_3_2_5]] ) # we verified the loss both for normalized and unnormalized targets for this one snake_case_ = torch.tensor([0.5_1_4_2] ) if config.norm_pix_loss else torch.tensor([0.6_4_6_9] ) elif model_name == "videomae-large": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.7_1_4_9, 0.7_9_9_7, 0.6_9_6_6], [0.6_7_6_8, 0.7_8_6_9, 0.6_9_4_8], [0.5_1_3_9, 0.6_2_2_1, 0.5_6_0_5]] ) elif model_name == "videomae-large-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.0_7_7_1, 0.0_0_1_1, -0.3_6_2_5] ) elif model_name == "videomae-huge-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.2_4_3_3, 0.1_6_3_2, -0.4_8_9_4] ) elif model_name == "videomae-base-short-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.6_5_8_8, 0.0_9_9_0, -0.2_4_9_3] ) elif model_name == "videomae-base-finetuned-kinetics": snake_case_ = torch.Size([1, 400] ) snake_case_ = torch.tensor([0.3_6_6_9, -0.0_6_8_8, -0.2_4_2_1] ) elif model_name == "videomae-base-short-ssv2": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.4_7_1_2, 0.5_2_9_6, 0.5_7_8_6], [0.2_2_7_8, 0.2_7_2_9, 0.4_0_2_6], [0.0_3_5_2, 0.0_7_3_0, 0.2_5_0_6]] ) elif model_name == "videomae-base-short-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([-0.0_5_3_7, -0.1_5_3_9, -0.3_2_6_6] ) elif model_name == "videomae-base-ssv2": snake_case_ = torch.Size([1, 1408, 1536] ) snake_case_ = torch.tensor([[0.8_1_3_1, 0.8_7_2_7, 0.8_5_4_6], [0.7_3_6_6, 0.9_3_7_7, 0.8_8_7_0], [0.5_9_3_5, 0.8_8_7_4, 0.8_5_6_4]] ) elif model_name == "videomae-base-finetuned-ssv2": snake_case_ = torch.Size([1, 174] ) snake_case_ = torch.tensor([0.1_9_6_1, -0.8_3_3_7, -0.6_3_8_9] ) else: raise ValueError(F"Model name not supported. Should be one of {model_names}" ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3], __UpperCAmelCase, atol=1e-4 ) else: print('''Logits:''', logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3], __UpperCAmelCase, atol=1e-4 ) print('''Logits ok!''' ) # verify loss, if applicable if model_name == "videomae-base-short": snake_case_ = outputs.loss assert torch.allclose(__UpperCAmelCase, __UpperCAmelCase, atol=1e-4 ) print('''Loss ok!''' ) if pytorch_dump_folder_path is not None: print(F"Saving model and image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__UpperCAmelCase ) model.save_pretrained(__UpperCAmelCase ) if push_to_hub: print('''Pushing to the hub...''' ) model.push_to_hub(__UpperCAmelCase, organization='''nielsr''' ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4', type=str, help=( 'URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct' ' download link.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default='/Users/nielsrogge/Documents/VideoMAE/Test', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--model_name', default='videomae-base', type=str, help='Name of the model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) a : Union[str, Any] = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
56
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_pytesseract, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class a : @staticmethod def A_ ( *lowercase_ : int , **lowercase_ : str ): pass @is_pipeline_test @require_vision @require_timm @require_torch class a ( unittest.TestCase ): snake_case_ = MODEL_FOR_OBJECT_DETECTION_MAPPING def A_ ( self : Any , lowercase_ : List[Any] , lowercase_ : Optional[int] , lowercase_ : List[str] ): snake_case_ = ObjectDetectionPipeline(model=lowercase_ , image_processor=lowercase_ ) return object_detector, ["./tests/fixtures/tests_samples/COCO/000000039769.png"] def A_ ( self : Dict , lowercase_ : List[Any] , lowercase_ : int ): snake_case_ = object_detector('''./tests/fixtures/tests_samples/COCO/000000039769.png''' , threshold=0.0 ) self.assertGreater(len(lowercase_ ) , 0 ) for detected_object in outputs: self.assertEqual( lowercase_ , { '''score''': ANY(lowercase_ ), '''label''': ANY(lowercase_ ), '''box''': {'''xmin''': ANY(lowercase_ ), '''ymin''': ANY(lowercase_ ), '''xmax''': ANY(lowercase_ ), '''ymax''': ANY(lowercase_ )}, } , ) import datasets snake_case_ = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) snake_case_ = [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] snake_case_ = object_detector(lowercase_ , threshold=0.0 ) self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) for outputs in batch_outputs: self.assertGreater(len(lowercase_ ) , 0 ) for detected_object in outputs: self.assertEqual( lowercase_ , { '''score''': ANY(lowercase_ ), '''label''': ANY(lowercase_ ), '''box''': {'''xmin''': ANY(lowercase_ ), '''ymin''': ANY(lowercase_ ), '''xmax''': ANY(lowercase_ ), '''ymax''': ANY(lowercase_ )}, } , ) @require_tf @unittest.skip('''Object detection not implemented in TF''' ) def A_ ( self : int ): pass @require_torch def A_ ( self : Tuple ): snake_case_ = '''hf-internal-testing/tiny-detr-mobilenetsv3''' snake_case_ = AutoModelForObjectDetection.from_pretrained(lowercase_ ) snake_case_ = AutoFeatureExtractor.from_pretrained(lowercase_ ) snake_case_ = ObjectDetectionPipeline(model=lowercase_ , feature_extractor=lowercase_ ) snake_case_ = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' , threshold=0.0 ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ] , ) snake_case_ = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] , threshold=0.0 , ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ [ {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ], [ {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, {'''score''': 0.3376, '''label''': '''LABEL_0''', '''box''': {'''xmin''': 159, '''ymin''': 120, '''xmax''': 480, '''ymax''': 359}}, ], ] , ) @require_torch @slow def A_ ( self : Optional[int] ): snake_case_ = '''facebook/detr-resnet-50''' snake_case_ = AutoModelForObjectDetection.from_pretrained(lowercase_ ) snake_case_ = AutoFeatureExtractor.from_pretrained(lowercase_ ) snake_case_ = ObjectDetectionPipeline(model=lowercase_ , feature_extractor=lowercase_ ) snake_case_ = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ] , ) snake_case_ = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ] , ) @require_torch @slow def A_ ( self : Tuple ): snake_case_ = '''facebook/detr-resnet-50''' snake_case_ = pipeline('''object-detection''' , model=lowercase_ ) snake_case_ = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ] , ) snake_case_ = object_detector( [ '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''http://images.cocodataset.org/val2017/000000039769.jpg''', ] ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], [ {'''score''': 0.9982, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 70, '''xmax''': 175, '''ymax''': 117}}, {'''score''': 0.9960, '''label''': '''remote''', '''box''': {'''xmin''': 333, '''ymin''': 72, '''xmax''': 368, '''ymax''': 187}}, {'''score''': 0.9955, '''label''': '''couch''', '''box''': {'''xmin''': 0, '''ymin''': 1, '''xmax''': 639, '''ymax''': 473}}, {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ], ] , ) @require_torch @slow def A_ ( self : str ): snake_case_ = 0.9985 snake_case_ = '''facebook/detr-resnet-50''' snake_case_ = pipeline('''object-detection''' , model=lowercase_ ) snake_case_ = object_detector('''http://images.cocodataset.org/val2017/000000039769.jpg''' , threshold=lowercase_ ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ {'''score''': 0.9988, '''label''': '''cat''', '''box''': {'''xmin''': 13, '''ymin''': 52, '''xmax''': 314, '''ymax''': 470}}, {'''score''': 0.9987, '''label''': '''cat''', '''box''': {'''xmin''': 345, '''ymin''': 23, '''xmax''': 640, '''ymax''': 368}}, ] , ) @require_torch @require_pytesseract @slow def A_ ( self : Dict ): snake_case_ = '''Narsil/layoutlmv3-finetuned-funsd''' snake_case_ = 0.9993 snake_case_ = pipeline('''object-detection''' , model=lowercase_ , threshold=lowercase_ ) snake_case_ = object_detector( '''https://huggingface.co/spaces/impira/docquery/resolve/2359223c1837a7587402bda0f2643382a6eefeab/invoice.png''' ) self.assertEqual( nested_simplify(lowercase_ , decimals=4 ) , [ {'''score''': 0.9993, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 294, '''ymin''': 254, '''xmax''': 343, '''ymax''': 264}}, {'''score''': 0.9993, '''label''': '''I-ANSWER''', '''box''': {'''xmin''': 294, '''ymin''': 254, '''xmax''': 343, '''ymax''': 264}}, ] , )
56
1
"""simple docstring""" import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class lowerCamelCase : '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=7 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=99 , _lowerCamelCase=64 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=512 , _lowerCamelCase=16 , _lowerCamelCase=2 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=4 , _lowerCamelCase=None , ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = parent UpperCAmelCase__ : int = batch_size UpperCAmelCase__ : Union[str, Any] = seq_length UpperCAmelCase__ : Optional[int] = is_training UpperCAmelCase__ : Optional[int] = use_input_mask UpperCAmelCase__ : int = use_token_type_ids UpperCAmelCase__ : Dict = use_labels UpperCAmelCase__ : Dict = vocab_size UpperCAmelCase__ : Optional[int] = hidden_size UpperCAmelCase__ : Union[str, Any] = num_hidden_layers UpperCAmelCase__ : Optional[int] = num_attention_heads UpperCAmelCase__ : str = intermediate_size UpperCAmelCase__ : int = hidden_act UpperCAmelCase__ : List[Any] = hidden_dropout_prob UpperCAmelCase__ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase__ : Tuple = max_position_embeddings UpperCAmelCase__ : List[str] = type_vocab_size UpperCAmelCase__ : Dict = type_sequence_label_size UpperCAmelCase__ : int = initializer_range UpperCAmelCase__ : Tuple = num_labels UpperCAmelCase__ : List[str] = num_choices UpperCAmelCase__ : Union[str, Any] = scope UpperCAmelCase__ : List[Any] = vocab_size - 1 def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ : Dict = None if self.use_input_mask: UpperCAmelCase__ : str = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : Optional[Any] = None if self.use_labels: UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ : Optional[Any] = self.get_config() return config, input_ids, input_mask, token_labels def _a (self ): """simple docstring""" return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def _a (self ): """simple docstring""" UpperCAmelCase__ : List[Any] = self.prepare_config_and_inputs() UpperCAmelCase__ : str = True return config, input_ids, input_mask, token_labels def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : List[str] = GPTNeoXModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Optional[int] = model(_lowerCamelCase , attention_mask=_lowerCamelCase ) UpperCAmelCase__ : Optional[Any] = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : str = True UpperCAmelCase__ : Union[str, Any] = GPTNeoXModel(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Tuple = model(_lowerCamelCase , attention_mask=_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Any = GPTNeoXForCausalLM(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Union[str, Any] = model(_lowerCamelCase , attention_mask=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : List[str] = self.num_labels UpperCAmelCase__ : Any = GPTNeoXForQuestionAnswering(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Any = model(_lowerCamelCase , attention_mask=_lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.num_labels UpperCAmelCase__ : Dict = GPTNeoXForSequenceClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : str = model(_lowerCamelCase , attention_mask=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Any = self.num_labels UpperCAmelCase__ : Dict = GPTNeoXForTokenClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : str = model(_lowerCamelCase , attention_mask=_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = True UpperCAmelCase__ : str = GPTNeoXForCausalLM(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() # first forward pass UpperCAmelCase__ : List[Any] = model(_lowerCamelCase , attention_mask=_lowerCamelCase , use_cache=_lowerCamelCase ) UpperCAmelCase__ : List[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase__ : str = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase__ : List[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and UpperCAmelCase__ : Optional[Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCAmelCase__ : Union[str, Any] = torch.cat([input_mask, next_mask] , dim=-1 ) UpperCAmelCase__ : List[Any] = model(_lowerCamelCase , attention_mask=_lowerCamelCase , output_hidden_states=_lowerCamelCase ) UpperCAmelCase__ : int = output_from_no_past["""hidden_states"""][0] UpperCAmelCase__ : Union[str, Any] = model( _lowerCamelCase , attention_mask=_lowerCamelCase , past_key_values=_lowerCamelCase , output_hidden_states=_lowerCamelCase , )["""hidden_states"""][0] # select random slice UpperCAmelCase__ : Optional[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCAmelCase__ : List[str] = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase__ : str = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_lowerCamelCase , _lowerCamelCase , atol=1e-3 ) ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Dict = self.prepare_config_and_inputs() UpperCAmelCase__ : Optional[int] = config_and_inputs UpperCAmelCase__ : List[str] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE = (GPTNeoXForCausalLM,) if is_torch_available() else () SCREAMING_SNAKE_CASE = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _a (self ): """simple docstring""" UpperCAmelCase__ : Any = GPTNeoXModelTester(self ) UpperCAmelCase__ : int = ConfigTester(self , config_class=_lowerCamelCase , hidden_size=64 , num_attention_heads=8 ) def _a (self ): """simple docstring""" self.config_tester.run_common_tests() def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() UpperCAmelCase__ : Optional[Any] = None self.model_tester.create_and_check_model_as_decoder(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCamelCase ) @unittest.skip(reason="""Feed forward chunking is not implemented""" ) def _a (self ): """simple docstring""" pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def _a (self , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : Optional[int] = ids_tensor([1, 10] , config.vocab_size ) UpperCAmelCase__ : int = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase__ : Tuple = GPTNeoXModel(_lowerCamelCase ) original_model.to(_lowerCamelCase ) original_model.eval() UpperCAmelCase__ : Optional[int] = original_model(_lowerCamelCase ).last_hidden_state UpperCAmelCase__ : Optional[Any] = original_model(_lowerCamelCase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase__ : str = {"""type""": scaling_type, """factor""": 10.0} UpperCAmelCase__ : Tuple = GPTNeoXModel(_lowerCamelCase ) scaled_model.to(_lowerCamelCase ) scaled_model.eval() UpperCAmelCase__ : Dict = scaled_model(_lowerCamelCase ).last_hidden_state UpperCAmelCase__ : Optional[int] = scaled_model(_lowerCamelCase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(_lowerCamelCase , _lowerCamelCase , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(_lowerCamelCase , _lowerCamelCase , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(_lowerCamelCase , _lowerCamelCase , atol=1e-5 ) ) @require_torch class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def _a (self ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = AutoTokenizer.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) for checkpointing in [True, False]: UpperCAmelCase__ : Any = GPTNeoXForCausalLM.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(_lowerCamelCase ) UpperCAmelCase__ : Any = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(_lowerCamelCase ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 UpperCAmelCase__ : Union[str, Any] = """My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure""" UpperCAmelCase__ : List[Any] = model.generate(**_lowerCamelCase , do_sample=_lowerCamelCase , max_new_tokens=20 ) UpperCAmelCase__ : Tuple = tokenizer.batch_decode(_lowerCamelCase )[0] self.assertEqual(_lowerCamelCase , _lowerCamelCase )
360
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = dataset UpperCAmelCase__ : Union[str, Any] = process UpperCAmelCase__ : List[Any] = params def __len__(self ): """simple docstring""" return len(self.dataset ) def __getitem__(self , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : int = self.dataset[i] UpperCAmelCase__ : Any = self.process(_lowerCamelCase , **self.params ) return processed class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=None ): """simple docstring""" UpperCAmelCase__ : Tuple = loader UpperCAmelCase__ : int = infer UpperCAmelCase__ : Optional[Any] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether UpperCAmelCase__ : Tuple = None UpperCAmelCase__ : Any = loader_batch_size # Internal bookkeeping UpperCAmelCase__ : Tuple = None UpperCAmelCase__ : Union[str, Any] = None def __len__(self ): """simple docstring""" return len(self.loader ) def __iter__(self ): """simple docstring""" UpperCAmelCase__ : List[str] = iter(self.loader ) return self def _a (self ): """simple docstring""" if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice UpperCAmelCase__ : Optional[int] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) UpperCAmelCase__ : List[str] = {} for k, element in self._loader_batch_data.items(): if isinstance(_lowerCamelCase , _lowerCamelCase ): # Convert ModelOutput to tuple first UpperCAmelCase__ : List[Any] = element.to_tuple() if isinstance(element[0] , torch.Tensor ): UpperCAmelCase__ : str = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): UpperCAmelCase__ : List[str] = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(_lowerCamelCase , _lowerCamelCase ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): UpperCAmelCase__ : Optional[Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): UpperCAmelCase__ : int = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around UpperCAmelCase__ : str = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers UpperCAmelCase__ : Tuple = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers UpperCAmelCase__ : Dict = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. UpperCAmelCase__ : Optional[Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 UpperCAmelCase__ : Union[str, Any] = self._loader_batch_data.__class__(_lowerCamelCase ) self._loader_batch_index += 1 return result def _a (self ): """simple docstring""" if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch UpperCAmelCase__ : str = next(self.iterator ) UpperCAmelCase__ : Union[str, Any] = self.infer(_lowerCamelCase , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(_lowerCamelCase , torch.Tensor ): UpperCAmelCase__ : List[Any] = processed else: UpperCAmelCase__ : List[str] = list(processed.keys() )[0] UpperCAmelCase__ : List[str] = processed[key] if isinstance(_lowerCamelCase , _lowerCamelCase ): UpperCAmelCase__ : Any = len(_lowerCamelCase ) else: UpperCAmelCase__ : Union[str, Any] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. UpperCAmelCase__ : Optional[int] = observed_batch_size # Setting internal index to unwrap the batch UpperCAmelCase__ : List[Any] = processed UpperCAmelCase__ : Optional[int] = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=None ): """simple docstring""" super().__init__(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def __iter__(self ): """simple docstring""" UpperCAmelCase__ : Tuple = iter(self.loader ) UpperCAmelCase__ : List[Any] = None return self def _a (self ): """simple docstring""" if self.subiterator is None: UpperCAmelCase__ : Optional[Any] = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item UpperCAmelCase__ : List[str] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators UpperCAmelCase__ : Optional[Any] = self.infer(next(self.iterator ) , **self.params ) UpperCAmelCase__ : List[str] = next(self.subiterator ) return processed class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def __iter__(self ): """simple docstring""" UpperCAmelCase__ : str = iter(self.loader ) return self def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = False UpperCAmelCase__ : List[str] = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: UpperCAmelCase__ : List[Any] = self.loader_batch_item() UpperCAmelCase__ : Dict = item.pop("""is_last""" ) accumulator.append(_lowerCamelCase ) if is_last: return accumulator while not is_last: UpperCAmelCase__ : List[str] = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(_lowerCamelCase , torch.Tensor ): UpperCAmelCase__ : Dict = processed else: UpperCAmelCase__ : List[Any] = list(processed.keys() )[0] UpperCAmelCase__ : List[Any] = processed[key] if isinstance(_lowerCamelCase , _lowerCamelCase ): UpperCAmelCase__ : int = len(_lowerCamelCase ) else: UpperCAmelCase__ : List[Any] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. UpperCAmelCase__ : str = observed_batch_size UpperCAmelCase__ : Union[str, Any] = processed UpperCAmelCase__ : List[Any] = 0 while self._loader_batch_index < self.loader_batch_size: UpperCAmelCase__ : Union[str, Any] = self.loader_batch_item() UpperCAmelCase__ : int = item.pop("""is_last""" ) accumulator.append(_lowerCamelCase ) if is_last: return accumulator else: UpperCAmelCase__ : Any = processed UpperCAmelCase__ : Any = item.pop("""is_last""" ) accumulator.append(_lowerCamelCase ) return accumulator class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : List[Any] = dataset UpperCAmelCase__ : Union[str, Any] = key def __len__(self ): """simple docstring""" return len(self.dataset ) def __getitem__(self , _lowerCamelCase ): """simple docstring""" return self.dataset[i][self.key] class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : int = dataset UpperCAmelCase__ : Any = keya UpperCAmelCase__ : str = keya def __len__(self ): """simple docstring""" return len(self.dataset ) def __getitem__(self , _lowerCamelCase ): """simple docstring""" return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
166
0
'''simple docstring''' def snake_case_ ( __SCREAMING_SNAKE_CASE : Optional[int] ): """simple docstring""" lowercase_ : List[str] = [0] * len(__SCREAMING_SNAKE_CASE ) lowercase_ : Union[str, Any] = [] lowercase_ : int = [] lowercase_ : Optional[int] = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__SCREAMING_SNAKE_CASE ) ): if indegree[i] == 0: queue.append(__SCREAMING_SNAKE_CASE ) while queue: lowercase_ : List[str] = queue.pop(0 ) cnt += 1 topo.append(__SCREAMING_SNAKE_CASE ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(__SCREAMING_SNAKE_CASE ) if cnt != len(__SCREAMING_SNAKE_CASE ): print('''Cycle exists''' ) else: print(__SCREAMING_SNAKE_CASE ) # Adjacency List of Graph _lowercase : Dict = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
93
"""simple docstring""" import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def UpperCamelCase ( _lowerCAmelCase : Dict, _lowerCAmelCase : int=(), _lowerCAmelCase : Union[str, Any]=None, _lowerCAmelCase : Union[str, Any]="no", _lowerCAmelCase : Optional[int]="29500" ) -> Any: _UpperCAmelCase : Any = False _UpperCAmelCase : Dict = False if any(key.startswith("""KAGGLE""" ) for key in os.environ.keys() ): _UpperCAmelCase : Union[str, Any] = True elif "IPython" in sys.modules: _UpperCAmelCase : Dict = """google.colab""" in str(sys.modules["""IPython"""].get_ipython() ) try: _UpperCAmelCase : int = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( f'''Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.''' ) if (in_colab or in_kaggle) and (os.environ.get("""TPU_NAME""", _lowerCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( """To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside """ """your training function. Restart your notebook and make sure no cells initializes an """ """`Accelerator`.""" ) if num_processes is None: _UpperCAmelCase : List[Any] = 8 _UpperCAmelCase : int = PrepareForLaunch(_lowerCAmelCase, distributed_type="""TPU""" ) print(f'''Launching a training on {num_processes} TPU cores.''' ) xmp.spawn(_lowerCAmelCase, args=_lowerCAmelCase, nprocs=_lowerCAmelCase, start_method="""fork""" ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print("""Launching training on one GPU.""" ) else: print("""Launching training on one CPU.""" ) function(*_lowerCAmelCase ) else: if num_processes is None: raise ValueError( """You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.""" ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( """To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized """ """inside your training function. Restart your notebook and make sure no cells initializes an """ """`Accelerator`.""" ) if torch.cuda.is_initialized(): raise ValueError( """To launch a multi-GPU training from your notebook, you need to avoid running any instruction """ """using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA """ """function.""" ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_lowerCAmelCase, master_addr="""127.0.01""", master_port=_lowerCAmelCase, mixed_precision=_lowerCAmelCase ): _UpperCAmelCase : Any = PrepareForLaunch(_lowerCAmelCase, distributed_type="""MULTI_GPU""" ) print(f'''Launching training on {num_processes} GPUs.''' ) try: start_processes(_lowerCAmelCase, args=_lowerCAmelCase, nprocs=_lowerCAmelCase, start_method="""fork""" ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( """CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. """ """This likely stems from an outside import causing issues once the `notebook_launcher()` is called. """ """Please review your imports and test them when running the `notebook_launcher()` to identify """ """which one is problematic.""" ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): _UpperCAmelCase : Union[str, Any] = """1""" print("""Launching training on MPS.""" ) elif torch.cuda.is_available(): print("""Launching training on one GPU.""" ) else: print("""Launching training on CPU.""" ) function(*_lowerCAmelCase ) def UpperCamelCase ( _lowerCAmelCase : Optional[Any], _lowerCAmelCase : List[str]=(), _lowerCAmelCase : Optional[int]=2 ) -> Tuple: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_lowerCAmelCase, master_addr="""127.0.01""", master_port="""29500""", accelerate_mixed_precision="""no""", accelerate_debug_rdv_file=tmp_file.name, accelerate_use_cpu="""yes""", ): _UpperCAmelCase : Tuple = PrepareForLaunch(_lowerCAmelCase, debug=_lowerCAmelCase ) start_processes(_lowerCAmelCase, args=_lowerCAmelCase, nprocs=_lowerCAmelCase, start_method="""fork""" )
246
0
from __future__ import annotations def __lowerCamelCase ( __magic_name__ : list[int] ): # This function is recursive a__: Optional[Any] =len(__magic_name__ ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else a__: Optional[Any] =array[0] a__: Dict =False a__: Optional[Any] =1 a__: list[int] =[] while not is_found and i < array_length: if array[i] < pivot: a__: List[Any] =True a__: List[Any] =[element for element in array[i:] if element >= array[i]] a__: Optional[int] =longest_subsequence(__magic_name__ ) if len(__magic_name__ ) > len(__magic_name__ ): a__: Dict =temp_array else: i += 1 a__: List[Any] =[element for element in array[1:] if element >= pivot] a__: int =[pivot, *longest_subsequence(__magic_name__ )] if len(__magic_name__ ) > len(__magic_name__ ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
366
def __lowerCamelCase ( __magic_name__ : int ): return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
42
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowercase = '''▁''' _lowercase = {'''vocab_file''': '''spiece.model'''} _lowercase = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''} } _lowercase = { '''google/pegasus-xsum''': 5_12, } _lowercase = logging.get_logger(__name__) class lowerCAmelCase_ ( _lowercase ): '''simple docstring''' _lowerCamelCase: List[str] = VOCAB_FILES_NAMES _lowerCamelCase: Optional[Any] = VOCAB_FILES_NAMES _lowerCamelCase: Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase: Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase: str = ['''input_ids''', '''attention_mask'''] def __init__( self : Union[str, Any] ,A_ : int ,A_ : List[Any]="<pad>" ,A_ : List[str]="</s>" ,A_ : Any="<unk>" ,A_ : Union[str, Any]="<mask_2>" ,A_ : int="<mask_1>" ,A_ : str=None ,A_ : List[Any]=103 ,A_ : Optional[Dict[str, Any]] = None ,**A_ : List[Any] ,) -> None: A = offset if additional_special_tokens is not None: if not isinstance(A_ ,A_ ): raise TypeError( F'additional_special_tokens should be of type {type(A_ )}, but is' F' {type(A_ )}' ) A = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F'<unk_{i}>' for i in range(len(A_ ) ,self.offset - 1 ) ] if len(set(A_ ) ) != len(A_ ): raise ValueError( 'Please make sure that the provided additional_special_tokens do not contain an incorrectly' F' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.' ) A = additional_special_tokens_extended else: A = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F'<unk_{i}>' for i in range(2 ,self.offset )] A = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=A_ ,unk_token=A_ ,mask_token=A_ ,pad_token=A_ ,mask_token_sent=A_ ,offset=A_ ,additional_special_tokens=A_ ,sp_model_kwargs=self.sp_model_kwargs ,**A_ ,) A = mask_token_sent A = vocab_file A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(A_ ) # add special tokens to encoder dict A = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, } ) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 ,self.offset - 1 )} ) A = {v: k for k, v in self.encoder.items()} @property def _SCREAMING_SNAKE_CASE ( self : str ) -> int: return len(self.sp_model ) + self.offset def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Dict[str, int]: A = {self.convert_ids_to_tokens(A_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Union[str, Any] ) -> Optional[int]: A = self.__dict__.copy() A = None return state def __setstate__( self : str ,A_ : Tuple ) -> List[str]: A = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): A = {} A = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _SCREAMING_SNAKE_CASE ( self : str ,A_ : str ) -> List[str]: return self.sp_model.encode(A_ ,out_type=A_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ,A_ : str ) -> int: if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] A = self.sp_model.piece_to_id(A_ ) return sp_id + self.offset def _SCREAMING_SNAKE_CASE ( self : List[Any] ,A_ : int ) -> str: if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: A = self.sp_model.IdToPiece(index - self.offset ) return token def _SCREAMING_SNAKE_CASE ( self : int ,A_ : int ) -> Any: A = [] A = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(A_ ) + token A = [] else: current_sub_tokens.append(A_ ) out_string += self.sp_model.decode(A_ ) return out_string.strip() def _SCREAMING_SNAKE_CASE ( self : Dict ,A_ : List[str]=False ) -> Optional[Any]: return 1 def _SCREAMING_SNAKE_CASE ( self : List[Any] ,A_ : Dict ) -> Optional[int]: A = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _SCREAMING_SNAKE_CASE ( self : Any ,A_ : List ,A_ : Optional[List] = None ,A_ : bool = False ) -> List[int]: if already_has_special_tokens: return self._special_token_mask(A_ ) elif token_ids_a is None: return self._special_token_mask(A_ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,A_ : Any ,A_ : Optional[int]=None ) -> List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _SCREAMING_SNAKE_CASE ( self : List[Any] ,A_ : str ,A_ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(A_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A = os.path.join( A_ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(A_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,A_ ) elif not os.path.isfile(self.vocab_file ): with open(A_ ,'wb' ) as fi: A = self.sp_model.serialized_model_proto() fi.write(A_ ) return (out_vocab_file,)
74
"""simple docstring""" import argparse import struct import unittest class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Tuple ,A_ : bytes ) -> None: A = data # Initialize hash values A = [ 0X6_A_0_9_E_6_6_7, 0XB_B_6_7_A_E_8_5, 0X3_C_6_E_F_3_7_2, 0XA_5_4_F_F_5_3_A, 0X5_1_0_E_5_2_7_F, 0X9_B_0_5_6_8_8_C, 0X1_F_8_3_D_9_A_B, 0X5_B_E_0_C_D_1_9, ] # Initialize round constants A = [ 0X4_2_8_A_2_F_9_8, 0X7_1_3_7_4_4_9_1, 0XB_5_C_0_F_B_C_F, 0XE_9_B_5_D_B_A_5, 0X3_9_5_6_C_2_5_B, 0X5_9_F_1_1_1_F_1, 0X9_2_3_F_8_2_A_4, 0XA_B_1_C_5_E_D_5, 0XD_8_0_7_A_A_9_8, 0X1_2_8_3_5_B_0_1, 0X2_4_3_1_8_5_B_E, 0X5_5_0_C_7_D_C_3, 0X7_2_B_E_5_D_7_4, 0X8_0_D_E_B_1_F_E, 0X9_B_D_C_0_6_A_7, 0XC_1_9_B_F_1_7_4, 0XE_4_9_B_6_9_C_1, 0XE_F_B_E_4_7_8_6, 0X0_F_C_1_9_D_C_6, 0X2_4_0_C_A_1_C_C, 0X2_D_E_9_2_C_6_F, 0X4_A_7_4_8_4_A_A, 0X5_C_B_0_A_9_D_C, 0X7_6_F_9_8_8_D_A, 0X9_8_3_E_5_1_5_2, 0XA_8_3_1_C_6_6_D, 0XB_0_0_3_2_7_C_8, 0XB_F_5_9_7_F_C_7, 0XC_6_E_0_0_B_F_3, 0XD_5_A_7_9_1_4_7, 0X0_6_C_A_6_3_5_1, 0X1_4_2_9_2_9_6_7, 0X2_7_B_7_0_A_8_5, 0X2_E_1_B_2_1_3_8, 0X4_D_2_C_6_D_F_C, 0X5_3_3_8_0_D_1_3, 0X6_5_0_A_7_3_5_4, 0X7_6_6_A_0_A_B_B, 0X8_1_C_2_C_9_2_E, 0X9_2_7_2_2_C_8_5, 0XA_2_B_F_E_8_A_1, 0XA_8_1_A_6_6_4_B, 0XC_2_4_B_8_B_7_0, 0XC_7_6_C_5_1_A_3, 0XD_1_9_2_E_8_1_9, 0XD_6_9_9_0_6_2_4, 0XF_4_0_E_3_5_8_5, 0X1_0_6_A_A_0_7_0, 0X1_9_A_4_C_1_1_6, 0X1_E_3_7_6_C_0_8, 0X2_7_4_8_7_7_4_C, 0X3_4_B_0_B_C_B_5, 0X3_9_1_C_0_C_B_3, 0X4_E_D_8_A_A_4_A, 0X5_B_9_C_C_A_4_F, 0X6_8_2_E_6_F_F_3, 0X7_4_8_F_8_2_E_E, 0X7_8_A_5_6_3_6_F, 0X8_4_C_8_7_8_1_4, 0X8_C_C_7_0_2_0_8, 0X9_0_B_E_F_F_F_A, 0XA_4_5_0_6_C_E_B, 0XB_E_F_9_A_3_F_7, 0XC_6_7_1_7_8_F_2, ] A = self.preprocessing(self.data ) self.final_hash() @staticmethod def _SCREAMING_SNAKE_CASE ( A_ : bytes ) -> bytes: A = B'\x80' + (B'\x00' * (63 - (len(A_ ) + 8) % 64)) A = struct.pack('>Q' ,(len(A_ ) * 8) ) return data + padding + big_endian_integer def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> None: # Convert into blocks of 64 bytes A = [ self.preprocessed_data[x : x + 64] for x in range(0 ,len(self.preprocessed_data ) ,64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers A = list(struct.unpack('>16L' ,A_ ) ) # add 48 0-ed integers words += [0] * 48 A , A , A , A , A , A , A , A = self.hashes for index in range(0 ,64 ): if index > 15: # modify the zero-ed indexes at the end of the array A = ( self.ror(words[index - 15] ,7 ) ^ self.ror(words[index - 15] ,18 ) ^ (words[index - 15] >> 3) ) A = ( self.ror(words[index - 2] ,17 ) ^ self.ror(words[index - 2] ,19 ) ^ (words[index - 2] >> 10) ) A = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X1_0_0_0_0_0_0_0_0 # Compression A = self.ror(A_ ,6 ) ^ self.ror(A_ ,11 ) ^ self.ror(A_ ,25 ) A = (e & f) ^ ((~e & 0XF_F_F_F_F_F_F_F) & g) A = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X1_0_0_0_0_0_0_0_0 A = self.ror(A_ ,2 ) ^ self.ror(A_ ,13 ) ^ self.ror(A_ ,22 ) A = (a & b) ^ (a & c) ^ (b & c) A = (sa + maj) % 0X1_0_0_0_0_0_0_0_0 A , A , A , A , A , A , A , A = ( g, f, e, ((d + tempa) % 0X1_0_0_0_0_0_0_0_0), c, b, a, ((tempa + tempa) % 0X1_0_0_0_0_0_0_0_0), ) A = [a, b, c, d, e, f, g, h] # Modify final values A = [ ((element + mutated_hash_values[index]) % 0X1_0_0_0_0_0_0_0_0) for index, element in enumerate(self.hashes ) ] A = ''.join([hex(A_ )[2:].zfill(8 ) for value in self.hashes] ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ,A_ : int ,A_ : int ) -> int: return 0XF_F_F_F_F_F_F_F & (value << (32 - rotations)) | (value >> rotations) class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> None: import hashlib A = bytes('Test String' ,'utf-8' ) self.assertEqual(SHAaaa(A_ ).hash ,hashlib.shaaaa(A_ ).hexdigest() ) def _snake_case ( ): import doctest doctest.testmod() A = argparse.ArgumentParser() parser.add_argument( '-s' , '--string' , dest='input_string' , default='Hello World!! Welcome to Cryptography' , help='Hash the string' , ) parser.add_argument( '-f' , '--file' , dest='input_file' , help='Hash contents of a file' ) A = parser.parse_args() A = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , 'rb' ) as f: A = f.read() else: A = bytes(snake_case__ , 'utf-8' ) print(SHAaaa(snake_case__ ).hash ) if __name__ == "__main__": main()
74
1
"""simple docstring""" from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig __UpperCamelCase : str = logging.get_logger(__name__) # General docstring __UpperCamelCase : Tuple = '''ResNetConfig''' # Base docstring __UpperCamelCase : str = '''microsoft/resnet-50''' __UpperCamelCase : int = [1, 2_0_4_8, 7, 7] # Image classification docstring __UpperCamelCase : Tuple = '''microsoft/resnet-50''' __UpperCamelCase : List[Any] = '''tiger cat''' __UpperCamelCase : List[Any] = [ '''microsoft/resnet-50''', # See all resnet models at https://huggingface.co/models?filter=resnet ] class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Union[str, Any] ,lowercase_ : int ,lowercase_ : int ,lowercase_ : int = 3 ,lowercase_ : int = 1 ,lowercase_ : str = "relu" ): super().__init__() lowerCAmelCase__ : Union[str, Any] = nn.Convad( lowercase_ ,lowercase_ ,kernel_size=lowercase_ ,stride=lowercase_ ,padding=kernel_size // 2 ,bias=lowercase_ ) lowerCAmelCase__ : str = nn.BatchNormad(lowercase_ ) lowerCAmelCase__ : Optional[Any] = ACTaFN[activation] if activation is not None else nn.Identity() def __lowerCAmelCase ( self : Optional[int] ,lowercase_ : Tensor ): lowerCAmelCase__ : Tuple = self.convolution(lowercase_ ) lowerCAmelCase__ : Union[str, Any] = self.normalization(lowercase_ ) lowerCAmelCase__ : Optional[Any] = self.activation(lowercase_ ) return hidden_state class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase_ : ResNetConfig ): super().__init__() lowerCAmelCase__ : Tuple = ResNetConvLayer( config.num_channels ,config.embedding_size ,kernel_size=7 ,stride=2 ,activation=config.hidden_act ) lowerCAmelCase__ : Optional[int] = nn.MaxPoolad(kernel_size=3 ,stride=2 ,padding=1 ) lowerCAmelCase__ : Dict = config.num_channels def __lowerCAmelCase ( self : int ,lowercase_ : Tensor ): lowerCAmelCase__ : Tuple = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) lowerCAmelCase__ : int = self.embedder(lowercase_ ) lowerCAmelCase__ : Optional[int] = self.pooler(lowercase_ ) return embedding class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : List[Any] ,lowercase_ : int ,lowercase_ : int ,lowercase_ : int = 2 ): super().__init__() lowerCAmelCase__ : int = nn.Convad(lowercase_ ,lowercase_ ,kernel_size=1 ,stride=lowercase_ ,bias=lowercase_ ) lowerCAmelCase__ : Union[str, Any] = nn.BatchNormad(lowercase_ ) def __lowerCAmelCase ( self : Optional[int] ,lowercase_ : Tensor ): lowerCAmelCase__ : str = self.convolution(lowercase_ ) lowerCAmelCase__ : Dict = self.normalization(lowercase_ ) return hidden_state class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Dict ,lowercase_ : int ,lowercase_ : int ,lowercase_ : int = 1 ,lowercase_ : str = "relu" ): super().__init__() lowerCAmelCase__ : Tuple = in_channels != out_channels or stride != 1 lowerCAmelCase__ : Any = ( ResNetShortCut(lowercase_ ,lowercase_ ,stride=lowercase_ ) if should_apply_shortcut else nn.Identity() ) lowerCAmelCase__ : List[str] = nn.Sequential( ResNetConvLayer(lowercase_ ,lowercase_ ,stride=lowercase_ ) ,ResNetConvLayer(lowercase_ ,lowercase_ ,activation=lowercase_ ) ,) lowerCAmelCase__ : int = ACTaFN[activation] def __lowerCAmelCase ( self : Optional[int] ,lowercase_ : Dict ): lowerCAmelCase__ : Dict = hidden_state lowerCAmelCase__ : Dict = self.layer(lowercase_ ) lowerCAmelCase__ : int = self.shortcut(lowercase_ ) hidden_state += residual lowerCAmelCase__ : Optional[int] = self.activation(lowercase_ ) return hidden_state class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Tuple ,lowercase_ : int ,lowercase_ : int ,lowercase_ : int = 1 ,lowercase_ : str = "relu" ,lowercase_ : int = 4 ): super().__init__() lowerCAmelCase__ : int = in_channels != out_channels or stride != 1 lowerCAmelCase__ : str = out_channels // reduction lowerCAmelCase__ : Union[str, Any] = ( ResNetShortCut(lowercase_ ,lowercase_ ,stride=lowercase_ ) if should_apply_shortcut else nn.Identity() ) lowerCAmelCase__ : Optional[int] = nn.Sequential( ResNetConvLayer(lowercase_ ,lowercase_ ,kernel_size=1 ) ,ResNetConvLayer(lowercase_ ,lowercase_ ,stride=lowercase_ ) ,ResNetConvLayer(lowercase_ ,lowercase_ ,kernel_size=1 ,activation=lowercase_ ) ,) lowerCAmelCase__ : Optional[Any] = ACTaFN[activation] def __lowerCAmelCase ( self : Any ,lowercase_ : Optional[int] ): lowerCAmelCase__ : str = hidden_state lowerCAmelCase__ : List[str] = self.layer(lowercase_ ) lowerCAmelCase__ : Any = self.shortcut(lowercase_ ) hidden_state += residual lowerCAmelCase__ : Dict = self.activation(lowercase_ ) return hidden_state class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Tuple ,lowercase_ : ResNetConfig ,lowercase_ : int ,lowercase_ : int ,lowercase_ : int = 2 ,lowercase_ : int = 2 ,): super().__init__() lowerCAmelCase__ : Optional[int] = ResNetBottleNeckLayer if config.layer_type == '''bottleneck''' else ResNetBasicLayer lowerCAmelCase__ : Union[str, Any] = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(lowercase_ ,lowercase_ ,stride=lowercase_ ,activation=config.hidden_act ) ,*[layer(lowercase_ ,lowercase_ ,activation=config.hidden_act ) for _ in range(depth - 1 )] ,) def __lowerCAmelCase ( self : int ,lowercase_ : Tensor ): lowerCAmelCase__ : Dict = input for layer in self.layers: lowerCAmelCase__ : Dict = layer(lowercase_ ) return hidden_state class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] ,lowercase_ : ResNetConfig ): super().__init__() lowerCAmelCase__ : int = nn.ModuleList([] ) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( lowercase_ ,config.embedding_size ,config.hidden_sizes[0] ,stride=2 if config.downsample_in_first_stage else 1 ,depth=config.depths[0] ,) ) lowerCAmelCase__ : Tuple = zip(config.hidden_sizes ,config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(lowercase_ ,config.depths[1:] ): self.stages.append(ResNetStage(lowercase_ ,lowercase_ ,lowercase_ ,depth=lowercase_ ) ) def __lowerCAmelCase ( self : Optional[Any] ,lowercase_ : Tensor ,lowercase_ : bool = False ,lowercase_ : bool = True ): lowerCAmelCase__ : Any = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowerCAmelCase__ : Dict = hidden_states + (hidden_state,) lowerCAmelCase__ : Any = stage_module(lowercase_ ) if output_hidden_states: lowerCAmelCase__ : str = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention( last_hidden_state=lowercase_ ,hidden_states=lowercase_ ,) class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = ResNetConfig lowercase__ = "resnet" lowercase__ = "pixel_values" lowercase__ = True def __lowerCAmelCase ( self : Optional[int] ,lowercase_ : int ): if isinstance(lowercase_ ,nn.Convad ): nn.init.kaiming_normal_(module.weight ,mode='''fan_out''' ,nonlinearity='''relu''' ) elif isinstance(lowercase_ ,(nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight ,1 ) nn.init.constant_(module.bias ,0 ) def __lowerCAmelCase ( self : Dict ,lowercase_ : Any ,lowercase_ : Optional[Any]=False ): if isinstance(lowercase_ ,lowercase_ ): lowerCAmelCase__ : Union[str, Any] = value __UpperCamelCase : Dict = R''' This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and behavior. Parameters: config ([`ResNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. ''' __UpperCamelCase : int = R''' Args: pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConvNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( "The bare ResNet model outputting raw features without any specific head on top." , a_ , ) class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" def __init__( self : int ,lowercase_ : Dict ): super().__init__(lowercase_ ) lowerCAmelCase__ : int = config lowerCAmelCase__ : Dict = ResNetEmbeddings(lowercase_ ) lowerCAmelCase__ : Tuple = ResNetEncoder(lowercase_ ) lowerCAmelCase__ : Dict = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowercase_ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC ,output_type=lowercase_ ,config_class=_CONFIG_FOR_DOC ,modality='''vision''' ,expected_output=_EXPECTED_OUTPUT_SHAPE ,) def __lowerCAmelCase ( self : List[Any] ,lowercase_ : Tensor ,lowercase_ : Optional[bool] = None ,lowercase_ : Optional[bool] = None ): lowerCAmelCase__ : Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCAmelCase__ : Any = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase__ : Optional[Any] = self.embedder(lowercase_ ) lowerCAmelCase__ : Optional[Any] = self.encoder( lowercase_ ,output_hidden_states=lowercase_ ,return_dict=lowercase_ ) lowerCAmelCase__ : int = encoder_outputs[0] lowerCAmelCase__ : List[Any] = self.pooler(lowercase_ ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowercase_ ,pooler_output=lowercase_ ,hidden_states=encoder_outputs.hidden_states ,) @add_start_docstrings( "\n ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n " , a_ , ) class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" def __init__( self : List[Any] ,lowercase_ : Any ): super().__init__(lowercase_ ) lowerCAmelCase__ : int = config.num_labels lowerCAmelCase__ : int = ResNetModel(lowercase_ ) # classification head lowerCAmelCase__ : Optional[int] = nn.Sequential( nn.Flatten() ,nn.Linear(config.hidden_sizes[-1] ,config.num_labels ) if config.num_labels > 0 else nn.Identity() ,) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowercase_ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT ,output_type=lowercase_ ,config_class=_CONFIG_FOR_DOC ,expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT ,) def __lowerCAmelCase ( self : Dict ,lowercase_ : Optional[torch.FloatTensor] = None ,lowercase_ : Optional[torch.LongTensor] = None ,lowercase_ : Optional[bool] = None ,lowercase_ : Optional[bool] = None ,): lowerCAmelCase__ : List[Any] = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase__ : List[str] = self.resnet(lowercase_ ,output_hidden_states=lowercase_ ,return_dict=lowercase_ ) lowerCAmelCase__ : int = outputs.pooler_output if return_dict else outputs[1] lowerCAmelCase__ : Optional[int] = self.classifier(lowercase_ ) lowerCAmelCase__ : Union[str, Any] = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowerCAmelCase__ : List[str] = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowerCAmelCase__ : int = '''single_label_classification''' else: lowerCAmelCase__ : int = '''multi_label_classification''' if self.config.problem_type == "regression": lowerCAmelCase__ : List[Any] = MSELoss() if self.num_labels == 1: lowerCAmelCase__ : List[Any] = loss_fct(logits.squeeze() ,labels.squeeze() ) else: lowerCAmelCase__ : Tuple = loss_fct(lowercase_ ,lowercase_ ) elif self.config.problem_type == "single_label_classification": lowerCAmelCase__ : List[Any] = CrossEntropyLoss() lowerCAmelCase__ : int = loss_fct(logits.view(-1 ,self.num_labels ) ,labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowerCAmelCase__ : List[Any] = BCEWithLogitsLoss() lowerCAmelCase__ : Dict = loss_fct(lowercase_ ,lowercase_ ) if not return_dict: lowerCAmelCase__ : str = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=lowercase_ ,logits=lowercase_ ,hidden_states=outputs.hidden_states ) @add_start_docstrings( "\n ResNet backbone, to be used with frameworks like DETR and MaskFormer.\n " , a_ , ) class SCREAMING_SNAKE_CASE ( a_ , a_ ): """simple docstring""" def __init__( self : Any ,lowercase_ : Any ): super().__init__(lowercase_ ) super()._init_backbone(lowercase_ ) lowerCAmelCase__ : Union[str, Any] = [config.embedding_size] + config.hidden_sizes lowerCAmelCase__ : int = ResNetEmbeddings(lowercase_ ) lowerCAmelCase__ : Tuple = ResNetEncoder(lowercase_ ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(lowercase_ ) @replace_return_docstrings(output_type=lowercase_ ,config_class=_CONFIG_FOR_DOC ) def __lowerCAmelCase ( self : int ,lowercase_ : Tensor ,lowercase_ : Optional[bool] = None ,lowercase_ : Optional[bool] = None ): lowerCAmelCase__ : Any = return_dict if return_dict is not None else self.config.use_return_dict lowerCAmelCase__ : Union[str, Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowerCAmelCase__ : Dict = self.embedder(lowercase_ ) lowerCAmelCase__ : Optional[int] = self.encoder(lowercase_ ,output_hidden_states=lowercase_ ,return_dict=lowercase_ ) lowerCAmelCase__ : Optional[Any] = outputs.hidden_states lowerCAmelCase__ : Optional[int] = () for idx, stage in enumerate(self.stage_names ): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: lowerCAmelCase__ : int = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=lowercase_ ,hidden_states=outputs.hidden_states if output_hidden_states else None ,attentions=lowercase_ ,)
74
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml __UpperCamelCase : Any = NewType('''DataClass''', Any) __UpperCamelCase : List[str] = NewType('''DataClassType''', Any) def __SCREAMING_SNAKE_CASE ( A_ ): if isinstance(A_ , A_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f'Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).' ) def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : int = {str(A_ ): choice for choice in choices} return lambda A_ : str_to_choice.get(A_ , A_ ) def __SCREAMING_SNAKE_CASE ( *, A_ = None , A_ = None , A_ = dataclasses.MISSING , A_ = dataclasses.MISSING , A_ = None , **A_ , ): if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls lowerCAmelCase__ : Dict = {} if aliases is not None: lowerCAmelCase__ : int = aliases if help is not None: lowerCAmelCase__ : Optional[int] = help return dataclasses.field(metadata=A_ , default=A_ , default_factory=A_ , **A_ ) class SCREAMING_SNAKE_CASE ( a_ ): """simple docstring""" lowercase__ = 42 def __init__( self : Dict ,lowercase_ : Union[DataClassType, Iterable[DataClassType]] ,**lowercase_ : str ): # To make the default appear when using --help if "formatter_class" not in kwargs: lowerCAmelCase__ : Tuple = ArgumentDefaultsHelpFormatter super().__init__(**lowercase_ ) if dataclasses.is_dataclass(lowercase_ ): lowerCAmelCase__ : Tuple = [dataclass_types] lowerCAmelCase__ : List[str] = list(lowercase_ ) for dtype in self.dataclass_types: self._add_dataclass_arguments(lowercase_ ) @staticmethod def __lowerCAmelCase ( lowercase_ : ArgumentParser ,lowercase_ : dataclasses.Field ): lowerCAmelCase__ : Dict = F'--{field.name}' lowerCAmelCase__ : List[str] = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type ,lowercase_ ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) lowerCAmelCase__ : List[str] = kwargs.pop('''aliases''' ,[] ) if isinstance(lowercase_ ,lowercase_ ): lowerCAmelCase__ : Optional[Any] = [aliases] lowerCAmelCase__ : Union[str, Any] = getattr(field.type ,'''__origin__''' ,field.type ) if origin_type is Union or (hasattr(lowercase_ ,'''UnionType''' ) and isinstance(lowercase_ ,types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(lowercase_ ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' F' Problem encountered in field \'{field.name}\'.' ) if type(lowercase_ ) not in field.type.__args__: # filter `str` in Union lowerCAmelCase__ : int = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] lowerCAmelCase__ : List[str] = getattr(field.type ,'''__origin__''' ,field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) lowerCAmelCase__ : Optional[int] = ( field.type.__args__[0] if isinstance(lowercase_ ,field.type.__args__[1] ) else field.type.__args__[1] ) lowerCAmelCase__ : Optional[Any] = getattr(field.type ,'''__origin__''' ,field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) lowerCAmelCase__ : List[Any] = {} if origin_type is Literal or (isinstance(field.type ,lowercase_ ) and issubclass(field.type ,lowercase_ )): if origin_type is Literal: lowerCAmelCase__ : Union[str, Any] = field.type.__args__ else: lowerCAmelCase__ : Optional[Any] = [x.value for x in field.type] lowerCAmelCase__ : List[str] = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: lowerCAmelCase__ : int = field.default else: lowerCAmelCase__ : Any = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument lowerCAmelCase__ : List[Any] = copy(lowercase_ ) # Hack because type=bool in argparse does not behave as we want. lowerCAmelCase__ : Tuple = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. lowerCAmelCase__ : List[Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way lowerCAmelCase__ : Tuple = default # This tells argparse we accept 0 or 1 value after --field_name lowerCAmelCase__ : Union[str, Any] = '''?''' # This is the value that will get picked if we do --field_name (without value) lowerCAmelCase__ : Any = True elif isclass(lowercase_ ) and issubclass(lowercase_ ,lowercase_ ): lowerCAmelCase__ : List[str] = field.type.__args__[0] lowerCAmelCase__ : str = '''+''' if field.default_factory is not dataclasses.MISSING: lowerCAmelCase__ : Dict = field.default_factory() elif field.default is dataclasses.MISSING: lowerCAmelCase__ : str = True else: lowerCAmelCase__ : List[Any] = field.type if field.default is not dataclasses.MISSING: lowerCAmelCase__ : str = field.default elif field.default_factory is not dataclasses.MISSING: lowerCAmelCase__ : Any = field.default_factory() else: lowerCAmelCase__ : Optional[Any] = True parser.add_argument(lowercase_ ,*lowercase_ ,**lowercase_ ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): lowerCAmelCase__ : Optional[Any] = False parser.add_argument(F'--no_{field.name}' ,action='''store_false''' ,dest=field.name ,**lowercase_ ) def __lowerCAmelCase ( self : str ,lowercase_ : DataClassType ): if hasattr(lowercase_ ,'''_argument_group_name''' ): lowerCAmelCase__ : Optional[int] = self.add_argument_group(dtype._argument_group_name ) else: lowerCAmelCase__ : List[str] = self try: lowerCAmelCase__ : Dict[str, type] = get_type_hints(lowercase_ ) except NameError: raise RuntimeError( F'Type resolution failed for {dtype}. Try declaring the class in global scope or ' '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 1_0) and "unsupported operand type(s) for |" in str(lowercase_ ): lowerCAmelCase__ : int = '''.'''.join(map(lowercase_ ,sys.version_info[:3] ) ) raise RuntimeError( F'Type resolution failed for {dtype} on Python {python_version}. Try removing ' '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(lowercase_ ): if not field.init: continue lowerCAmelCase__ : Any = type_hints[field.name] self._parse_dataclass_field(lowercase_ ,lowercase_ ) def __lowerCAmelCase ( self : Any ,lowercase_ : Optional[Any]=None ,lowercase_ : str=False ,lowercase_ : str=True ,lowercase_ : Any=None ,lowercase_ : List[str]=None ,): if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): lowerCAmelCase__ : int = [] if args_filename: args_files.append(Path(lowercase_ ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values lowerCAmelCase__ : List[str] = ArgumentParser() args_file_parser.add_argument(lowercase_ ,type=lowercase_ ,action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) lowerCAmelCase__ ,lowerCAmelCase__ : List[str] = args_file_parser.parse_known_args(args=lowercase_ ) lowerCAmelCase__ : int = vars(lowercase_ ).get(args_file_flag.lstrip('''-''' ) ,lowercase_ ) if cmd_args_file_paths: args_files.extend([Path(lowercase_ ) for p in cmd_args_file_paths] ) lowerCAmelCase__ : Tuple = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last lowerCAmelCase__ : Dict = file_args + args if args is not None else file_args + sys.argv[1:] lowerCAmelCase__ ,lowerCAmelCase__ : Optional[Any] = self.parse_known_args(args=lowercase_ ) lowerCAmelCase__ : Optional[Any] = [] for dtype in self.dataclass_types: lowerCAmelCase__ : int = {f.name for f in dataclasses.fields(lowercase_ ) if f.init} lowerCAmelCase__ : int = {k: v for k, v in vars(lowercase_ ).items() if k in keys} for k in keys: delattr(lowercase_ ,lowercase_ ) lowerCAmelCase__ : Optional[Any] = dtype(**lowercase_ ) outputs.append(lowercase_ ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(lowercase_ ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F'Some specified arguments are not used by the HfArgumentParser: {remaining_args}' ) return (*outputs,) def __lowerCAmelCase ( self : Any ,lowercase_ : Dict[str, Any] ,lowercase_ : bool = False ): lowerCAmelCase__ : List[Any] = set(args.keys() ) lowerCAmelCase__ : Any = [] for dtype in self.dataclass_types: lowerCAmelCase__ : Optional[Any] = {f.name for f in dataclasses.fields(lowercase_ ) if f.init} lowerCAmelCase__ : Union[str, Any] = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) lowerCAmelCase__ : Union[str, Any] = dtype(**lowercase_ ) outputs.append(lowercase_ ) if not allow_extra_keys and unused_keys: raise ValueError(F'Some keys are not used by the HfArgumentParser: {sorted(lowercase_ )}' ) return tuple(lowercase_ ) def __lowerCAmelCase ( self : Optional[int] ,lowercase_ : str ,lowercase_ : bool = False ): with open(Path(lowercase_ ) ,encoding='''utf-8''' ) as open_json_file: lowerCAmelCase__ : Union[str, Any] = json.loads(open_json_file.read() ) lowerCAmelCase__ : List[str] = self.parse_dict(lowercase_ ,allow_extra_keys=lowercase_ ) return tuple(lowercase_ ) def __lowerCAmelCase ( self : Dict ,lowercase_ : str ,lowercase_ : bool = False ): lowerCAmelCase__ : Tuple = self.parse_dict(yaml.safe_load(Path(lowercase_ ).read_text() ) ,allow_extra_keys=lowercase_ ) return tuple(lowercase_ )
74
1
"""simple docstring""" class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Any: _a = name _a = value _a = weight def __repr__( self ) -> Tuple: return F'{self.__class__.__name__}({self.name}, {self.value}, {self.weight})' def _UpperCAmelCase ( self ) -> str: return self.value def _UpperCAmelCase ( self ) -> Tuple: return self.name def _UpperCAmelCase ( self ) -> Union[str, Any]: return self.weight def _UpperCAmelCase ( self ) -> Union[str, Any]: return self.value / self.weight def A_ ( _lowerCAmelCase : int, _lowerCAmelCase : str, _lowerCAmelCase : Union[str, Any] ): """simple docstring""" _a = [] for i in range(len(_lowerCAmelCase ) ): menu.append(Things(name[i], value[i], weight[i] ) ) return menu def A_ ( _lowerCAmelCase : Tuple, _lowerCAmelCase : Dict, _lowerCAmelCase : int ): """simple docstring""" _a = sorted(_lowerCAmelCase, key=_lowerCAmelCase, reverse=_lowerCAmelCase ) _a = [] _a = 0.0, 0.0 for i in range(len(_lowerCAmelCase ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def A_ ( ): """simple docstring""" pass if __name__ == "__main__": import doctest doctest.testmod()
320
from scipy.stats import pearsonr import datasets _lowerCamelCase =""" Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ _lowerCamelCase =""" Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ _lowerCamelCase =""" @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class A__ ( datasets.Metric): def UpperCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) , reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"""] , ) def UpperCamelCase__ ( self , __magic_name__ , __magic_name__ , __magic_name__=False ): if return_pvalue: lowerCamelCase : Optional[Any] = pearsonr(__magic_name__ , __magic_name__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__magic_name__ , __magic_name__ )[0] )}
287
0
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : float , __UpperCamelCase : float , __UpperCamelCase : int ) -> float: if principal <= 0: raise Exception('''Principal borrowed must be > 0''' ) if rate_per_annum < 0: raise Exception('''Rate of interest must be >= 0''' ) if years_to_repay <= 0 or not isinstance(__UpperCamelCase , __UpperCamelCase ): raise Exception('''Years to repay must be an integer > 0''' ) # Yearly rate is divided by 12 to get monthly rate UpperCAmelCase_ = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly UpperCAmelCase_ = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
361
def SCREAMING_SNAKE_CASE ( __UpperCamelCase : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase_ = len(__UpperCamelCase ) while cur > 1: # Find the maximum number in arr UpperCAmelCase_ = arr.index(max(arr[0:cur] ) ) # Reverse from 0 to mi UpperCAmelCase_ = arr[mi::-1] + arr[mi + 1 : len(__UpperCamelCase )] # Reverse whole list UpperCAmelCase_ = arr[cur - 1 :: -1] + arr[cur : len(__UpperCamelCase )] cur -= 1 return arr if __name__ == "__main__": _lowerCamelCase = input('Enter numbers separated by a comma:\n').strip() _lowerCamelCase = [int(item) for item in user_input.split(',')] print(pancake_sort(unsorted))
177
0
'''simple docstring''' def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : list[list[int]] ): def update_area_of_max_square(_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int ) -> int: # BASE CASE if row >= rows or col >= cols: return 0 __a : Dict = update_area_of_max_square(_SCREAMING_SNAKE_CASE , col + 1 ) __a : Dict = update_area_of_max_square(row + 1 , col + 1 ) __a : Optional[Any] = update_area_of_max_square(row + 1 , _SCREAMING_SNAKE_CASE ) if mat[row][col]: __a : str = 1 + min([right, diagonal, down] ) __a : int = max(largest_square_area[0] , _SCREAMING_SNAKE_CASE ) return sub_problem_sol else: return 0 __a : Any = [0] update_area_of_max_square(0 , 0 ) return largest_square_area[0] def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : list[list[int]] ): def update_area_of_max_square_using_dp_array( _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : list[list[int]] ) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] __a : Tuple = update_area_of_max_square_using_dp_array(_SCREAMING_SNAKE_CASE , col + 1 , _SCREAMING_SNAKE_CASE ) __a : Optional[int] = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , _SCREAMING_SNAKE_CASE ) __a : Any = update_area_of_max_square_using_dp_array(row + 1 , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if mat[row][col]: __a : Union[str, Any] = 1 + min([right, diagonal, down] ) __a : Any = max(largest_square_area[0] , _SCREAMING_SNAKE_CASE ) __a : Dict = sub_problem_sol return sub_problem_sol else: return 0 __a : List[str] = [0] __a : List[Any] = [[-1] * cols for _ in range(_SCREAMING_SNAKE_CASE )] update_area_of_max_square_using_dp_array(0 , 0 , _SCREAMING_SNAKE_CASE ) return largest_square_area[0] def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : list[list[int]] ): __a : List[str] = [[0] * (cols + 1) for _ in range(rows + 1 )] __a : List[Any] = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __a : Optional[Any] = dp_array[row][col + 1] __a : Union[str, Any] = dp_array[row + 1][col + 1] __a : int = dp_array[row + 1][col] if mat[row][col] == 1: __a : Tuple = 1 + min(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __a : int = max(dp_array[row][col] , _SCREAMING_SNAKE_CASE ) else: __a : Union[str, Any] = 0 return largest_square_area def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : list[list[int]] ): __a : Tuple = [0] * (cols + 1) __a : Dict = [0] * (cols + 1) __a : Optional[int] = 0 for row in range(rows - 1 , -1 , -1 ): for col in range(cols - 1 , -1 , -1 ): __a : Optional[Any] = current_row[col + 1] __a : Union[str, Any] = next_row[col + 1] __a : Dict = next_row[col] if mat[row][col] == 1: __a : int = 1 + min(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __a : Union[str, Any] = max(current_row[col] , _SCREAMING_SNAKE_CASE ) else: __a : Union[str, Any] = 0 __a : List[str] = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
27
'''simple docstring''' import unittest import numpy as np from diffusers import OnnxStableDiffusionInpaintPipelineLegacy from diffusers.utils.testing_utils import ( is_onnx_available, load_image, load_numpy, nightly, require_onnxruntime, require_torch_gpu, ) if is_onnx_available(): import onnxruntime as ort @nightly @require_onnxruntime @require_torch_gpu class __UpperCamelCase ( unittest.TestCase ): @property def __UpperCAmelCase ( self ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = ort.SessionOptions() __a : Dict = False return options def __UpperCAmelCase ( self ): '''simple docstring''' __a : Tuple = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo.png' ) __a : int = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/overture-creations-5sI6fQgYIuo_mask.png' ) __a : Dict = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/in_paint/red_cat_sitting_on_a_park_bench_onnx.npy' ) # using the PNDM scheduler by default __a : str = OnnxStableDiffusionInpaintPipelineLegacy.from_pretrained( 'CompVis/stable-diffusion-v1-4' , revision='onnx' , safety_checker=__a , feature_extractor=__a , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__a ) __a : Tuple = 'A red cat sitting on a park bench' __a : int = np.random.RandomState(0 ) __a : Tuple = pipe( prompt=__a , image=__a , mask_image=__a , strength=0.75 , guidance_scale=7.5 , num_inference_steps=15 , generator=__a , output_type='np' , ) __a : Tuple = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 1E-2
27
1
def _lowercase ( UpperCamelCase_ , UpperCamelCase_ ) -> int: '''simple docstring''' assert x is not None assert y is not None SCREAMING_SNAKE_CASE__ = len(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ = len(__lowerCAmelCase ) # declaring the array for storing the dp values SCREAMING_SNAKE_CASE__ = [[0] * (n + 1) for _ in range(m + 1 )] # noqa: E741 for i in range(1 , m + 1 ): for j in range(1 , n + 1 ): SCREAMING_SNAKE_CASE__ = 1 if x[i - 1] == y[j - 1] else 0 SCREAMING_SNAKE_CASE__ = max(l[i - 1][j] , l[i][j - 1] , l[i - 1][j - 1] + match ) SCREAMING_SNAKE_CASE__ = '' SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = m, n while i > 0 and j > 0: SCREAMING_SNAKE_CASE__ = 1 if x[i - 1] == y[j - 1] else 0 if l[i][j] == l[i - 1][j - 1] + match: if match == 1: SCREAMING_SNAKE_CASE__ = x[i - 1] + seq i -= 1 j -= 1 elif l[i][j] == l[i - 1][j]: i -= 1 else: j -= 1 return l[m][n], seq if __name__ == "__main__": __snake_case = """AGGTAB""" __snake_case = """GXTXAYB""" __snake_case = 4 __snake_case = """GTAB""" __snake_case ,__snake_case = longest_common_subsequence(a, b) print("""len =""", ln, """, sub-sequence =""", subseq) import doctest doctest.testmod()
353
import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_weights from .dataclasses import BnbQuantizationConfig from .modeling import ( find_tied_parameters, get_balanced_memory, infer_auto_device_map, load_checkpoint_in_model, offload_weight, set_module_tensor_to_device, ) if is_bnb_available(): import bitsandbytes as bnb from copy import deepcopy __snake_case = logging.getLogger(__name__) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = None , UpperCamelCase_ = False , ) -> Tuple: '''simple docstring''' SCREAMING_SNAKE_CASE__ = bnb_quantization_config.load_in_abit SCREAMING_SNAKE_CASE__ = bnb_quantization_config.load_in_abit if load_in_abit and not is_abit_bnb_available(): raise ImportError( 'You have a version of `bitsandbytes` that is not compatible with 8bit quantization,' ' make sure you have the latest version of `bitsandbytes` installed.' ) if load_in_abit and not is_abit_bnb_available(): raise ValueError( 'You have a version of `bitsandbytes` that is not compatible with 4bit quantization,' 'make sure you have the latest version of `bitsandbytes` installed.' ) SCREAMING_SNAKE_CASE__ = [] # custom device map if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and len(device_map.keys() ) > 1: SCREAMING_SNAKE_CASE__ = [key for key, value in device_map.items() if value in ['disk', 'cpu']] # We keep some modules such as the lm_head in their original dtype for numerical stability reasons if bnb_quantization_config.skip_modules is None: SCREAMING_SNAKE_CASE__ = get_keys_to_not_convert(UpperCamelCase_ ) # add cpu modules to skip modules only for 4-bit modules if load_in_abit: bnb_quantization_config.skip_modules.extend(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = bnb_quantization_config.skip_modules # We add the modules we want to keep in full precision if bnb_quantization_config.keep_in_fpaa_modules is None: SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = bnb_quantization_config.keep_in_fpaa_modules modules_to_not_convert.extend(UpperCamelCase_ ) # compatibility with peft SCREAMING_SNAKE_CASE__ = load_in_abit SCREAMING_SNAKE_CASE__ = load_in_abit SCREAMING_SNAKE_CASE__ = get_parameter_device(UpperCamelCase_ ) if model_device.type != "meta": # quantization of an already loaded model logger.warning( 'It is not recommended to quantize a loaded model. ' 'The model should be instantiated under the `init_empty_weights` context manager.' ) SCREAMING_SNAKE_CASE__ = replace_with_bnb_layers(UpperCamelCase_ , UpperCamelCase_ , modules_to_not_convert=UpperCamelCase_ ) # convert param to the right dtype SCREAMING_SNAKE_CASE__ = bnb_quantization_config.torch_dtype for name, param in model.state_dict().items(): if any(module_to_keep_in_fpaa in name for module_to_keep_in_fpaa in keep_in_fpaa_modules ): param.to(torch.floataa ) if param.dtype != torch.floataa: SCREAMING_SNAKE_CASE__ = name.replace('.weight' , '' ).replace('.bias' , '' ) SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if param is not None: param.to(torch.floataa ) elif torch.is_floating_point(UpperCamelCase_ ): param.to(UpperCamelCase_ ) if model_device.type == "cuda": # move everything to cpu in the first place because we can't do quantization if the weights are already on cuda model.cuda(torch.cuda.current_device() ) torch.cuda.empty_cache() elif torch.cuda.is_available(): model.to(torch.cuda.current_device() ) else: raise RuntimeError('No GPU found. A GPU is needed for quantization.' ) logger.info( F'The model device type is {model_device.type}. However, cuda is needed for quantization.' 'We move the model to cuda.' ) return model elif weights_location is None: raise RuntimeError( F'`weights_location` needs to be the folder path containing the weights of the model, but we found {weights_location} ' ) else: with init_empty_weights(): SCREAMING_SNAKE_CASE__ = replace_with_bnb_layers( UpperCamelCase_ , UpperCamelCase_ , modules_to_not_convert=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = get_quantized_model_device_map( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , max_memory=UpperCamelCase_ , no_split_module_classes=UpperCamelCase_ , ) if offload_state_dict is None and device_map is not None and "disk" in device_map.values(): SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = any(x in list(device_map.values() ) for x in ['cpu', 'disk'] ) load_checkpoint_in_model( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , dtype=bnb_quantization_config.torch_dtype , offload_folder=UpperCamelCase_ , offload_state_dict=UpperCamelCase_ , keep_in_fpaa_modules=bnb_quantization_config.keep_in_fpaa_modules , offload_abit_bnb=load_in_abit and offload , ) return dispatch_model(UpperCamelCase_ , device_map=UpperCamelCase_ , offload_dir=UpperCamelCase_ ) def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_=None ) -> Optional[Any]: '''simple docstring''' if device_map is None: if torch.cuda.is_available(): SCREAMING_SNAKE_CASE__ = {'': torch.cuda.current_device()} else: raise RuntimeError('No GPU found. A GPU is needed for quantization.' ) logger.info('The device_map was not initialized.' 'Setting device_map to `{\'\':torch.cuda.current_device()}`.' ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ): if device_map not in ["auto", "balanced", "balanced_low_0", "sequential"]: raise ValueError( 'If passing a string for `device_map`, please choose \'auto\', \'balanced\', \'balanced_low_0\' or ' '\'sequential\'.' ) SCREAMING_SNAKE_CASE__ = {} special_dtypes.update( { name: bnb_quantization_config.torch_dtype for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.skip_modules ) } ) special_dtypes.update( { name: torch.floataa for name, _ in model.named_parameters() if any(m in name for m in bnb_quantization_config.keep_in_fpaa_modules ) } ) SCREAMING_SNAKE_CASE__ = {} SCREAMING_SNAKE_CASE__ = special_dtypes SCREAMING_SNAKE_CASE__ = no_split_module_classes SCREAMING_SNAKE_CASE__ = bnb_quantization_config.target_dtype # get max_memory for each device. if device_map != "sequential": SCREAMING_SNAKE_CASE__ = get_balanced_memory( UpperCamelCase_ , low_zero=(device_map == 'balanced_low_0') , max_memory=UpperCamelCase_ , **UpperCamelCase_ , ) SCREAMING_SNAKE_CASE__ = max_memory SCREAMING_SNAKE_CASE__ = infer_auto_device_map(UpperCamelCase_ , **UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ): # check if don't have any quantized module on the cpu SCREAMING_SNAKE_CASE__ = bnb_quantization_config.skip_modules + bnb_quantization_config.keep_in_fpaa_modules SCREAMING_SNAKE_CASE__ = { key: device_map[key] for key in device_map.keys() if key not in modules_not_to_convert } for device in ["cpu", "disk"]: if device in device_map_without_some_modules.values(): if bnb_quantization_config.load_in_abit: raise ValueError( '\n Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit\n the quantized model. If you want to dispatch the model on the CPU or the disk while keeping\n these modules in `torch_dtype`, you need to pass a custom `device_map` to\n `load_and_quantize_model`. Check\n https://huggingface.co/docs/accelerate/main/en/usage_guides/quantization#offload-modules-to-cpu-and-disk\n for more details.\n ' ) else: logger.info( 'Some modules are are offloaded to the CPU or the disk. Note that these modules will be converted to 8-bit' ) del device_map_without_some_modules return device_map def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None , UpperCamelCase_=None ) -> Optional[Any]: '''simple docstring''' if modules_to_not_convert is None: SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = _replace_with_bnb_layers( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) if not has_been_replaced: logger.warning( 'You are loading your model in 8bit or 4bit but no linear modules were found in your model.' ' this can happen for some architectures such as gpt2 that uses Conv1D instead of Linear layers.' ' Please double check your model architecture, or submit an issue on github if you think this is' ' a bug.' ) return model def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_=None , UpperCamelCase_=None , ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = False for name, module in model.named_children(): if current_key_name is None: SCREAMING_SNAKE_CASE__ = [] current_key_name.append(UpperCamelCase_ ) if isinstance(UpperCamelCase_ , nn.Linear ) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` SCREAMING_SNAKE_CASE__ = '.'.join(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = True for key in modules_to_not_convert: if ( (key in current_key_name_str) and (key + "." in current_key_name_str) ) or key == current_key_name_str: SCREAMING_SNAKE_CASE__ = False break if proceed: # Load bnb module with empty weight and replace ``nn.Linear` module if bnb_quantization_config.load_in_abit: SCREAMING_SNAKE_CASE__ = bnb.nn.LinearabitLt( module.in_features , module.out_features , module.bias is not None , has_fpaa_weights=UpperCamelCase_ , threshold=bnb_quantization_config.llm_inta_threshold , ) elif bnb_quantization_config.load_in_abit: SCREAMING_SNAKE_CASE__ = bnb.nn.Linearabit( module.in_features , module.out_features , module.bias is not None , bnb_quantization_config.bnb_abit_compute_dtype , compress_statistics=bnb_quantization_config.bnb_abit_use_double_quant , quant_type=bnb_quantization_config.bnb_abit_quant_type , ) else: raise ValueError('load_in_8bit and load_in_4bit can\'t be both False' ) SCREAMING_SNAKE_CASE__ = module.weight.data if module.bias is not None: SCREAMING_SNAKE_CASE__ = module.bias.data bnb_module.requires_grad_(UpperCamelCase_ ) setattr(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = True if len(list(module.children() ) ) > 0: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = _replace_with_bnb_layers( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = has_been_replaced | _has_been_replaced # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def _lowercase ( UpperCamelCase_ ) -> Union[str, Any]: '''simple docstring''' with init_empty_weights(): SCREAMING_SNAKE_CASE__ = deepcopy(UpperCamelCase_ ) # this has 0 cost since it is done inside `init_empty_weights` context manager` SCREAMING_SNAKE_CASE__ = find_tied_parameters(UpperCamelCase_ ) # For compatibility with Accelerate < 0.18 if isinstance(UpperCamelCase_ , UpperCamelCase_ ): SCREAMING_SNAKE_CASE__ = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: SCREAMING_SNAKE_CASE__ = sum(UpperCamelCase_ , [] ) SCREAMING_SNAKE_CASE__ = len(UpperCamelCase_ ) > 0 # Check if it is a base model SCREAMING_SNAKE_CASE__ = False if hasattr(UpperCamelCase_ , 'base_model_prefix' ): SCREAMING_SNAKE_CASE__ = not hasattr(UpperCamelCase_ , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head SCREAMING_SNAKE_CASE__ = list(model.named_children() ) SCREAMING_SNAKE_CASE__ = [list_modules[-1][0]] # add last module together with tied weights SCREAMING_SNAKE_CASE__ = set(UpperCamelCase_ ) - set(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = list(set(UpperCamelCase_ ) ) + list(UpperCamelCase_ ) # remove ".weight" from the keys SCREAMING_SNAKE_CASE__ = ['.weight', '.bias'] SCREAMING_SNAKE_CASE__ = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: SCREAMING_SNAKE_CASE__ = name.replace(UpperCamelCase_ , '' ) filtered_module_names.append(UpperCamelCase_ ) return filtered_module_names def _lowercase ( UpperCamelCase_ ) -> str: '''simple docstring''' for m in model.modules(): if isinstance(UpperCamelCase_ , bnb.nn.Linearabit ): return True return False def _lowercase ( UpperCamelCase_ ) -> str: '''simple docstring''' return next(parameter.parameters() ).device def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> List[str]: '''simple docstring''' if fpaa_statistics is None: set_module_tensor_to_device(UpperCamelCase_ , UpperCamelCase_ , 0 , dtype=UpperCamelCase_ , value=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = param_name SCREAMING_SNAKE_CASE__ = model if "." in tensor_name: SCREAMING_SNAKE_CASE__ = tensor_name.split('.' ) for split in splits[:-1]: SCREAMING_SNAKE_CASE__ = getattr(UpperCamelCase_ , UpperCamelCase_ ) if new_module is None: raise ValueError(F'{module} has no attribute {split}.' ) SCREAMING_SNAKE_CASE__ = new_module SCREAMING_SNAKE_CASE__ = splits[-1] # offload weights SCREAMING_SNAKE_CASE__ = False offload_weight(module._parameters[tensor_name] , UpperCamelCase_ , UpperCamelCase_ , index=UpperCamelCase_ ) if hasattr(module._parameters[tensor_name] , 'SCB' ): offload_weight( module._parameters[tensor_name].SCB , param_name.replace('weight' , 'SCB' ) , UpperCamelCase_ , index=UpperCamelCase_ , ) else: offload_weight(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , index=UpperCamelCase_ ) offload_weight(UpperCamelCase_ , param_name.replace('weight' , 'SCB' ) , UpperCamelCase_ , index=UpperCamelCase_ ) set_module_tensor_to_device(UpperCamelCase_ , UpperCamelCase_ , 'meta' , dtype=UpperCamelCase_ , value=torch.empty(*param.size() ) )
169
0
def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: return x if y == 0 else greatest_common_divisor(_SCREAMING_SNAKE_CASE ,x % y ) def A ( _SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) -> int: return (x * y) // greatest_common_divisor(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) def A ( _SCREAMING_SNAKE_CASE = 20 ) -> int: lowerCamelCase : List[Any] = 1 for i in range(1 ,n + 1 ): lowerCamelCase : List[str] = lcm(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) return g if __name__ == "__main__": print(f'''{solution() = }''')
48
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def a_ ( lowerCAmelCase_ : Optional[int] ): __lowerCAmelCase = filter(lambda lowerCAmelCase_ : p.requires_grad, model.parameters() ) __lowerCAmelCase = sum([np.prod(p.size() ) for p in model_parameters] ) return params _snake_case : Dict = logging.getLogger(__name__) def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Optional[int] ): if metric == "rouge2": __lowerCAmelCase = '{val_avg_rouge2:.4f}-{step_count}' elif metric == "bleu": __lowerCAmelCase = '{val_avg_bleu:.4f}-{step_count}' elif metric == "em": __lowerCAmelCase = '{val_avg_em:.4f}-{step_count}' else: raise NotImplementedError( F"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ' function.' ) __lowerCAmelCase = ModelCheckpoint( dirpath=lowerCAmelCase_, filename=lowerCAmelCase_, monitor=F"""val_{metric}""", mode='max', save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def a_ ( lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Any ): return EarlyStopping( monitor=F"""val_{metric}""", mode='min' if 'loss' in metric else 'max', patience=lowerCAmelCase_, verbose=lowerCAmelCase_, ) class _UpperCAmelCase ( pl.Callback ): """simple docstring""" def lowercase ( self : Tuple , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : int ) -> Any: __lowerCAmelCase = {f"""lr_group_{i}""": param['lr'] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Optional[int] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule , lowerCAmelCase_ : str , lowerCAmelCase_ : List[Any]=True ) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __lowerCAmelCase = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['log', 'progress_bar', 'preds']} ) # Log results __lowerCAmelCase = Path(pl_module.hparams.output_dir ) if type_path == "test": __lowerCAmelCase = od / 'test_results.txt' __lowerCAmelCase = od / 'test_generations.txt' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __lowerCAmelCase = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __lowerCAmelCase = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) generations_file.parent.mkdir(exist_ok=lowerCAmelCase_ ) with open(lowerCAmelCase_ , 'a+' ) as writer: for key in sorted(lowerCAmelCase_ ): if key in ["log", "progress_bar", "preds"]: continue __lowerCAmelCase = metrics[key] if isinstance(lowerCAmelCase_ , torch.Tensor ): __lowerCAmelCase = val.item() __lowerCAmelCase = f"""{key}: {val:.6f}\n""" writer.write(lowerCAmelCase_ ) if not save_generations: return if "preds" in metrics: __lowerCAmelCase = '\n'.join(metrics['preds'] ) generations_file.open('w+' ).write(lowerCAmelCase_ ) @rank_zero_only def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str] ) -> Dict: try: __lowerCAmelCase = pl_module.model.model.num_parameters() except AttributeError: __lowerCAmelCase = pl_module.model.num_parameters() __lowerCAmelCase = count_trainable_parameters(lowerCAmelCase_ ) # mp stands for million parameters trainer.logger.log_metrics({'n_params': npars, 'mp': npars / 1e6, 'grad_mp': n_trainable_pars / 1e6} ) @rank_zero_only def lowercase ( self : int , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : pl.LightningModule ) -> Any: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(lowerCAmelCase_ , lowerCAmelCase_ , 'test' ) @rank_zero_only def lowercase ( self : List[Any] , lowerCAmelCase_ : pl.Trainer , lowerCAmelCase_ : Any ) -> int: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
284
0
from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) _snake_case = 299792458 # Symbols _snake_case , _snake_case , _snake_case , _snake_case = symbols("ct x y z") def lowerCAmelCase_ ( snake_case_ ): if velocity > c: raise ValueError("""Speed must not exceed light speed 299,792,458 [m/s]!""" ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError("""Speed must be greater than or equal to 1!""" ) return velocity / c def lowerCAmelCase_ ( snake_case_ ): return 1 / sqrt(1 - beta(snake_case_ ) ** 2 ) def lowerCAmelCase_ ( snake_case_ ): return np.array( [ [gamma(snake_case_ ), -gamma(snake_case_ ) * beta(snake_case_ ), 0, 0], [-gamma(snake_case_ ) * beta(snake_case_ ), gamma(snake_case_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowerCAmelCase_ ( snake_case_,snake_case_ = None ): # Ensure event is not empty if event is None: _A : Dict = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(snake_case_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: _snake_case = transform(29979245) print("Example of four vector: ") print(f"""ct' = {four_vector[0]}""") print(f"""x' = {four_vector[1]}""") print(f"""y' = {four_vector[2]}""") print(f"""z' = {four_vector[3]}""") # Substitute symbols with numerical values _snake_case = {ct: c, x: 1, y: 1, z: 1} _snake_case = [four_vector[i].subs(sub_dict) for i in range(4)] print(f"""\n{numerical_vector}""")
343
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _snake_case = logging.get_logger(__name__) _snake_case = { "microsoft/resnet-50": "https://huggingface.co/microsoft/resnet-50/blob/main/config.json", } class lowercase ( UpperCamelCase__,UpperCamelCase__ ): _a = "resnet" _a = ["basic", "bottleneck"] def __init__( self , _a=3 , _a=64 , _a=[256, 512, 1024, 2048] , _a=[3, 4, 6, 3] , _a="bottleneck" , _a="relu" , _a=False , _a=None , _a=None , **_a , ) -> int: super().__init__(**_a ) if layer_type not in self.layer_types: raise ValueError(F'''layer_type={layer_type} is not one of {",".join(self.layer_types )}''' ) _A : Optional[Any] = num_channels _A : List[Any] = embedding_size _A : int = hidden_sizes _A : Union[str, Any] = depths _A : Optional[int] = layer_type _A : Any = hidden_act _A : List[Any] = downsample_in_first_stage _A : int = ["""stem"""] + [F'''stage{idx}''' for idx in range(1 , len(_a ) + 1 )] _A , _A : str = get_aligned_output_features_output_indices( out_features=_a , out_indices=_a , stage_names=self.stage_names ) class lowercase ( UpperCamelCase__ ): _a = version.parse("1.11" ) @property def a__ ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def a__ ( self ) -> float: return 1e-3
343
1
'''simple docstring''' from functools import lru_cache @lru_cache def snake_case__ ( _A: int ) -> int: '''simple docstring''' if num < 0: raise ValueError("""Number should not be negative.""" ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
272
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class a__( unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Dict = ViTImageProcessor if is_vision_available() else None @property def a_ ( self): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a_ ( self): """simple docstring""" lowerCAmelCase = (3, 32, 128) lowerCAmelCase = tempfile.mkdtemp() # fmt: off lowerCAmelCase = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on lowerCAmelCase = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase)))) lowerCAmelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""]) with open(self.vocab_file , """w""" , encoding="""utf-8""") as fp: fp.write(json.dumps(__lowerCAmelCase) + """\n""") lowerCAmelCase = { """do_normalize""": False, """do_resize""": True, """image_processor_type""": """ViTImageProcessor""", """resample""": 3, """size""": {"""height""": 32, """width""": 128}, } lowerCAmelCase = os.path.join(self.tmpdirname , __lowerCAmelCase) with open(self.image_processor_file , """w""" , encoding="""utf-8""") as fp: json.dump(__lowerCAmelCase , __lowerCAmelCase) def a_ ( self , **__lowerCAmelCase): """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__lowerCAmelCase) def a_ ( self , **__lowerCAmelCase): """simple docstring""" return ViTImageProcessor.from_pretrained(self.tmpdirname , **__lowerCAmelCase) def a_ ( self): """simple docstring""" shutil.rmtree(self.tmpdirname) def a_ ( self): """simple docstring""" lowerCAmelCase = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta) lowerCAmelCase = Image.fromarray(np.moveaxis(__lowerCAmelCase , 0 , -1)) return image_input def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = self.get_image_processor() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) processor.save_pretrained(self.tmpdirname) lowerCAmelCase = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=__lowerCAmelCase) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab()) self.assertIsInstance(processor.char_tokenizer , __lowerCAmelCase) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string()) self.assertIsInstance(processor.image_processor , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = self.get_image_processor() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) processor.save_pretrained(self.tmpdirname) lowerCAmelCase = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""") lowerCAmelCase = self.get_image_processor(do_normalize=__lowerCAmelCase , padding_value=1.0) lowerCAmelCase = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__lowerCAmelCase , padding_value=1.0) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.char_tokenizer , __lowerCAmelCase) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = self.prepare_image_inputs() lowerCAmelCase = image_processor(__lowerCAmelCase , return_tensors="""np""") lowerCAmelCase = processor(images=__lowerCAmelCase , return_tensors="""np""") for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = """test""" lowerCAmelCase = processor(text=__lowerCAmelCase) lowerCAmelCase = tokenizer(__lowerCAmelCase) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = """test""" lowerCAmelCase = self.prepare_image_inputs() lowerCAmelCase = processor(text=__lowerCAmelCase , images=__lowerCAmelCase) self.assertListEqual(list(inputs.keys()) , ["""pixel_values""", """labels"""]) # test if it raises when no input is passed with pytest.raises(__lowerCAmelCase): processor() def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] lowerCAmelCase = processor.char_decode(__lowerCAmelCase) lowerCAmelCase = tokenizer.batch_decode(__lowerCAmelCase) lowerCAmelCase = [seq.replace(""" """ , """""") for seq in decoded_tok] self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = None lowerCAmelCase = self.prepare_image_inputs() lowerCAmelCase = processor(text=__lowerCAmelCase , images=__lowerCAmelCase) self.assertListEqual(list(inputs.keys()) , processor.model_input_names) def a_ ( self): """simple docstring""" lowerCAmelCase = self.get_image_processor() lowerCAmelCase = self.get_tokenizer() lowerCAmelCase = MgpstrProcessor(tokenizer=__lowerCAmelCase , image_processor=__lowerCAmelCase) lowerCAmelCase = torch.randn(1 , 27 , 38) lowerCAmelCase = torch.randn(1 , 27 , 50257) lowerCAmelCase = torch.randn(1 , 27 , 30522) lowerCAmelCase = processor.batch_decode([char_input, bpe_input, wp_input]) self.assertListEqual(list(results.keys()) , ["""generated_text""", """scores""", """char_preds""", """bpe_preds""", """wp_preds"""])
272
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Tuple = {"configuration_wavlm": ["WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "WavLMConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : Union[str, Any] = [ "WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST", "WavLMForAudioFrameClassification", "WavLMForCTC", "WavLMForSequenceClassification", "WavLMForXVector", "WavLMModel", "WavLMPreTrainedModel", ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys _UpperCAmelCase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
158
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, **lowerCamelCase ): lowercase :List[Any] = AutoConfig.from_pretrained(lowerCamelCase, **lowerCamelCase ) lowercase :Union[str, Any] = AutoModelForSeqaSeqLM.from_config(lowerCamelCase ) model.save_pretrained(lowerCamelCase ) AutoTokenizer.from_pretrained(lowerCamelCase ).save_pretrained(lowerCamelCase ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
158
1
"""simple docstring""" from collections.abc import Callable import numpy as np def snake_case_ ( A_ : str, A_ : Any, A_ : List[Any], A_ : Any, A_ : Optional[int] ): '''simple docstring''' _lowerCamelCase : Any = int(np.ceil((x_end - xa) / step_size ) ) _lowerCamelCase : Optional[Any] = np.zeros((n + 1,) ) _lowerCamelCase : Any = ya _lowerCamelCase : Any = xa for k in range(__UpperCamelCase ): _lowerCamelCase : List[Any] = y[k] + step_size * ode_func(__UpperCamelCase, y[k] ) _lowerCamelCase : Union[str, Any] = y[k] + ( (step_size / 2) * (ode_func(__UpperCamelCase, y[k] ) + ode_func(x + step_size, __UpperCamelCase )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
72
def a__ ( __UpperCamelCase ): if n == 1 or not isinstance(__UpperCamelCase , __UpperCamelCase ): return 0 elif n == 2: return 1 else: SCREAMING_SNAKE_CASE_ = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def a__ ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = 2 while digits < n: index += 1 SCREAMING_SNAKE_CASE_ = len(str(fibonacci(__UpperCamelCase ) ) ) return index def a__ ( __UpperCamelCase = 1_0_0_0 ): return fibonacci_digits_index(__UpperCamelCase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
118
0
import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) _snake_case = logging.getLogger(__name__) _snake_case = '''Hello world! cécé herlolip''' _snake_case = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _UpperCamelCase ( snake_case__, snake_case__ ) -> int: __UpperCAmelCase : List[Any] = BertAbsConfig( temp_dir=".", finetune_bert=snake_case__, large=snake_case__, share_emb=snake_case__, use_bert_emb=snake_case__, encoder="bert", max_pos=512, enc_layers=6, enc_hidden_size=512, enc_heads=8, enc_ff_size=512, enc_dropout=0.2, dec_layers=6, dec_hidden_size=768, dec_heads=8, dec_ff_size=2048, dec_dropout=0.2, ) __UpperCAmelCase : str = torch.load(snake_case__, lambda snake_case__, snake_case__ : storage ) __UpperCAmelCase : str = AbsSummarizer(snake_case__, torch.device("cpu" ), snake_case__ ) original.eval() __UpperCAmelCase : Optional[Any] = BertAbsSummarizer(snake_case__, torch.device("cpu" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("convert the model" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("Make sure that the models' outputs are identical" ) __UpperCAmelCase : Any = BertTokenizer.from_pretrained("bert-base-uncased" ) # prepare the model inputs __UpperCAmelCase : int = tokenizer.encode("This is sample éàalj'-." ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(snake_case__ )) ) __UpperCAmelCase : str = torch.tensor(snake_case__ ).unsqueeze(0 ) __UpperCAmelCase : Optional[int] = tokenizer.encode("This is sample 3 éàalj'-." ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(snake_case__ )) ) __UpperCAmelCase : Any = torch.tensor(snake_case__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __UpperCAmelCase : str = encoder_input_ids __UpperCAmelCase : Dict = decoder_input_ids __UpperCAmelCase : List[str] = None __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[Any] = None __UpperCAmelCase : Dict = None __UpperCAmelCase : List[str] = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __UpperCAmelCase : Optional[int] = original(snake_case__, snake_case__, snake_case__, snake_case__, snake_case__, snake_case__, snake_case__ )[0] __UpperCAmelCase : str = original.generator(snake_case__ ) __UpperCAmelCase : Dict = new_model( snake_case__, snake_case__, snake_case__, snake_case__, snake_case__ )[0] __UpperCAmelCase : Any = new_model.generator(snake_case__ ) __UpperCAmelCase : int = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(snake_case__ ) ) __UpperCAmelCase : List[Any] = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(snake_case__ ) ) __UpperCAmelCase : Union[str, Any] = torch.allclose(snake_case__, snake_case__, atol=1e-3 ) if are_identical: logging.info("all weights are equal up to 1e-3" ) else: raise ValueError("the weights are different. The new model is likely different from the original one." ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("saving the model's state dictionary" ) torch.save( new_model.state_dict(), "./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) _snake_case = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
342
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) def _UpperCamelCase ( snake_case__ ) -> int: __UpperCAmelCase : int = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: __UpperCAmelCase : int = [144, 192, 240] __UpperCAmelCase : Optional[Any] = [16, 32, 64, 96, 128, 160, 640] elif "mobilevit_xs" in mobilevit_name: __UpperCAmelCase : Optional[Any] = [96, 120, 144] __UpperCAmelCase : Tuple = [16, 32, 48, 64, 80, 96, 384] elif "mobilevit_xxs" in mobilevit_name: __UpperCAmelCase : str = [64, 80, 96] __UpperCAmelCase : Optional[Any] = [16, 16, 24, 48, 64, 80, 320] __UpperCAmelCase : Tuple = 0.05 __UpperCAmelCase : Dict = 2.0 if mobilevit_name.startswith("deeplabv3_" ): __UpperCAmelCase : str = 512 __UpperCAmelCase : Any = 16 __UpperCAmelCase : str = 21 __UpperCAmelCase : Union[str, Any] = "pascal-voc-id2label.json" else: __UpperCAmelCase : Optional[Any] = 1000 __UpperCAmelCase : int = "imagenet-1k-id2label.json" __UpperCAmelCase : Dict = "huggingface/label-files" __UpperCAmelCase : int = json.load(open(hf_hub_download(snake_case__, snake_case__, repo_type="dataset" ), "r" ) ) __UpperCAmelCase : Any = {int(snake_case__ ): v for k, v in idalabel.items()} __UpperCAmelCase : int = idalabel __UpperCAmelCase : List[str] = {v: k for k, v in idalabel.items()} return config def _UpperCamelCase ( snake_case__, snake_case__=False ) -> Tuple: for i in range(1, 6 ): if f'''layer_{i}.''' in name: __UpperCAmelCase : Tuple = name.replace(f'''layer_{i}.''', f'''encoder.layer.{i - 1}.''' ) if "conv_1." in name: __UpperCAmelCase : Dict = name.replace("conv_1.", "conv_stem." ) if ".block." in name: __UpperCAmelCase : Optional[int] = name.replace(".block.", "." ) if "exp_1x1" in name: __UpperCAmelCase : Tuple = name.replace("exp_1x1", "expand_1x1" ) if "red_1x1" in name: __UpperCAmelCase : Optional[Any] = name.replace("red_1x1", "reduce_1x1" ) if ".local_rep.conv_3x3." in name: __UpperCAmelCase : Optional[int] = name.replace(".local_rep.conv_3x3.", ".conv_kxk." ) if ".local_rep.conv_1x1." in name: __UpperCAmelCase : Any = name.replace(".local_rep.conv_1x1.", ".conv_1x1." ) if ".norm." in name: __UpperCAmelCase : Dict = name.replace(".norm.", ".normalization." ) if ".conv." in name: __UpperCAmelCase : List[Any] = name.replace(".conv.", ".convolution." ) if ".conv_proj." in name: __UpperCAmelCase : List[str] = name.replace(".conv_proj.", ".conv_projection." ) for i in range(0, 2 ): for j in range(0, 4 ): if f'''.{i}.{j}.''' in name: __UpperCAmelCase : List[Any] = name.replace(f'''.{i}.{j}.''', f'''.{i}.layer.{j}.''' ) for i in range(2, 6 ): for j in range(0, 4 ): if f'''.{i}.{j}.''' in name: __UpperCAmelCase : Any = name.replace(f'''.{i}.{j}.''', f'''.{i}.''' ) if "expand_1x1" in name: __UpperCAmelCase : Optional[int] = name.replace("expand_1x1", "downsampling_layer.expand_1x1" ) if "conv_3x3" in name: __UpperCAmelCase : List[Any] = name.replace("conv_3x3", "downsampling_layer.conv_3x3" ) if "reduce_1x1" in name: __UpperCAmelCase : Dict = name.replace("reduce_1x1", "downsampling_layer.reduce_1x1" ) for i in range(2, 5 ): if f'''.global_rep.{i}.weight''' in name: __UpperCAmelCase : Any = name.replace(f'''.global_rep.{i}.weight''', ".layernorm.weight" ) if f'''.global_rep.{i}.bias''' in name: __UpperCAmelCase : Optional[Any] = name.replace(f'''.global_rep.{i}.bias''', ".layernorm.bias" ) if ".global_rep." in name: __UpperCAmelCase : Tuple = name.replace(".global_rep.", ".transformer." ) if ".pre_norm_mha.0." in name: __UpperCAmelCase : Optional[Any] = name.replace(".pre_norm_mha.0.", ".layernorm_before." ) if ".pre_norm_mha.1.out_proj." in name: __UpperCAmelCase : Tuple = name.replace(".pre_norm_mha.1.out_proj.", ".attention.output.dense." ) if ".pre_norm_ffn.0." in name: __UpperCAmelCase : Optional[Any] = name.replace(".pre_norm_ffn.0.", ".layernorm_after." ) if ".pre_norm_ffn.1." in name: __UpperCAmelCase : Dict = name.replace(".pre_norm_ffn.1.", ".intermediate.dense." ) if ".pre_norm_ffn.4." in name: __UpperCAmelCase : int = name.replace(".pre_norm_ffn.4.", ".output.dense." ) if ".transformer." in name: __UpperCAmelCase : Tuple = name.replace(".transformer.", ".transformer.layer." ) if ".aspp_layer." in name: __UpperCAmelCase : Any = name.replace(".aspp_layer.", "." ) if ".aspp_pool." in name: __UpperCAmelCase : Optional[Any] = name.replace(".aspp_pool.", "." ) if "seg_head." in name: __UpperCAmelCase : Optional[int] = name.replace("seg_head.", "segmentation_head." ) if "segmentation_head.classifier.classifier." in name: __UpperCAmelCase : str = name.replace("segmentation_head.classifier.classifier.", "segmentation_head.classifier." ) if "classifier.fc." in name: __UpperCAmelCase : Optional[Any] = name.replace("classifier.fc.", "classifier." ) elif (not base_model) and ("segmentation_head." not in name): __UpperCAmelCase : List[str] = "mobilevit." + name return name def _UpperCamelCase ( snake_case__, snake_case__, snake_case__=False ) -> Union[str, Any]: if base_model: __UpperCAmelCase : Optional[int] = "" else: __UpperCAmelCase : Tuple = "mobilevit." for key in orig_state_dict.copy().keys(): __UpperCAmelCase : Optional[int] = orig_state_dict.pop(snake_case__ ) if key[:8] == "encoder.": __UpperCAmelCase : str = key[8:] if "qkv" in key: __UpperCAmelCase : Tuple = key.split("." ) __UpperCAmelCase : List[Any] = int(key_split[0][6:] ) - 1 __UpperCAmelCase : Optional[Any] = int(key_split[3] ) __UpperCAmelCase : Tuple = model.get_submodule(f'''{model_prefix}encoder.layer.{layer_num}''' ) __UpperCAmelCase : List[str] = layer.transformer.layer[transformer_num].attention.attention.all_head_size __UpperCAmelCase : Optional[Any] = ( f'''{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.''' ) if "weight" in key: __UpperCAmelCase : Any = val[:dim, :] __UpperCAmelCase : Any = val[dim : dim * 2, :] __UpperCAmelCase : List[Any] = val[-dim:, :] else: __UpperCAmelCase : List[str] = val[:dim] __UpperCAmelCase : Optional[Any] = val[dim : dim * 2] __UpperCAmelCase : List[Any] = val[-dim:] else: __UpperCAmelCase : str = val return orig_state_dict def _UpperCamelCase ( ) -> Any: __UpperCAmelCase : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" __UpperCAmelCase : List[str] = Image.open(requests.get(snake_case__, stream=snake_case__ ).raw ) return im @torch.no_grad() def _UpperCamelCase ( snake_case__, snake_case__, snake_case__, snake_case__=False ) -> Optional[Any]: __UpperCAmelCase : Tuple = get_mobilevit_config(snake_case__ ) # load original state_dict __UpperCAmelCase : str = torch.load(snake_case__, map_location="cpu" ) # load 🤗 model if mobilevit_name.startswith("deeplabv3_" ): __UpperCAmelCase : Optional[int] = MobileViTForSemanticSegmentation(snake_case__ ).eval() else: __UpperCAmelCase : List[Any] = MobileViTForImageClassification(snake_case__ ).eval() __UpperCAmelCase : Dict = convert_state_dict(snake_case__, snake_case__ ) model.load_state_dict(snake_case__ ) # Check outputs on an image, prepared by MobileViTImageProcessor __UpperCAmelCase : Optional[Any] = MobileViTImageProcessor(crop_size=config.image_size, size=config.image_size + 32 ) __UpperCAmelCase : Any = image_processor(images=prepare_img(), return_tensors="pt" ) __UpperCAmelCase : Dict = model(**snake_case__ ) __UpperCAmelCase : Tuple = outputs.logits if mobilevit_name.startswith("deeplabv3_" ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": __UpperCAmelCase : int = torch.tensor( [ [[6.2065, 6.1292, 6.2070], [6.1079, 6.1254, 6.1747], [6.0042, 6.1071, 6.1034]], [[-6.9253, -6.8653, -7.0398], [-7.3218, -7.3983, -7.3670], [-7.1961, -7.2482, -7.1569]], [[-4.4723, -4.4348, -4.3769], [-5.3629, -5.4632, -5.4598], [-5.1587, -5.3402, -5.5059]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": __UpperCAmelCase : Tuple = torch.tensor( [ [[5.4449, 5.5733, 5.6314], [5.1815, 5.3930, 5.5963], [5.1656, 5.4333, 5.4853]], [[-9.4423, -9.7766, -9.6714], [-9.1581, -9.5720, -9.5519], [-9.1006, -9.6458, -9.5703]], [[-7.7721, -7.3716, -7.1583], [-8.4599, -8.0624, -7.7944], [-8.4172, -7.8366, -7.5025]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": __UpperCAmelCase : Any = torch.tensor( [ [[6.9811, 6.9743, 7.3123], [7.1777, 7.1931, 7.3938], [7.5633, 7.8050, 7.8901]], [[-10.5536, -10.2332, -10.2924], [-10.2336, -9.8624, -9.5964], [-10.8840, -10.8158, -10.6659]], [[-3.4938, -3.0631, -2.8620], [-3.4205, -2.8135, -2.6875], [-3.4179, -2.7945, -2.8750]], ] ) else: raise ValueError(f'''Unknown mobilevit_name: {mobilevit_name}''' ) assert torch.allclose(logits[0, :3, :3, :3], snake_case__, atol=1e-4 ) else: assert logits.shape == (1, 1000) if mobilevit_name == "mobilevit_s": __UpperCAmelCase : str = torch.tensor([-0.9866, 0.2392, -1.1241] ) elif mobilevit_name == "mobilevit_xs": __UpperCAmelCase : Tuple = torch.tensor([-2.4761, -0.9399, -1.9587] ) elif mobilevit_name == "mobilevit_xxs": __UpperCAmelCase : Union[str, Any] = torch.tensor([-1.9364, -1.2327, -0.4653] ) else: raise ValueError(f'''Unknown mobilevit_name: {mobilevit_name}''' ) assert torch.allclose(logits[0, :3], snake_case__, atol=1e-4 ) Path(snake_case__ ).mkdir(exist_ok=snake_case__ ) print(f'''Saving model {mobilevit_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(snake_case__ ) print(f'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(snake_case__ ) if push_to_hub: __UpperCAmelCase : List[str] = { "mobilevit_s": "mobilevit-small", "mobilevit_xs": "mobilevit-x-small", "mobilevit_xxs": "mobilevit-xx-small", "deeplabv3_mobilevit_s": "deeplabv3-mobilevit-small", "deeplabv3_mobilevit_xs": "deeplabv3-mobilevit-x-small", "deeplabv3_mobilevit_xxs": "deeplabv3-mobilevit-xx-small", } print("Pushing to the hub..." ) __UpperCAmelCase : int = model_mapping[mobilevit_name] image_processor.push_to_hub(snake_case__, organization="apple" ) model.push_to_hub(snake_case__, organization="apple" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--mobilevit_name''', default='''mobilevit_s''', type=str, help=( '''Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\',''' ''' \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'.''' ), ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original state dict (.pt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _snake_case = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
342
1
import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def a ( snake_case__: str ): '''simple docstring''' return (data["data"], data["target"]) def a ( snake_case__: Dict , snake_case__: int , snake_case__: Optional[Any] ): '''simple docstring''' lowercase_ = XGBRegressor(verbosity=0 , random_state=42 ) xgb.fit(_UpperCamelCase , _UpperCamelCase ) # Predict target for test data lowercase_ = xgb.predict(_UpperCamelCase ) lowercase_ = predictions.reshape(len(_UpperCamelCase ) , 1 ) return predictions def a ( ): '''simple docstring''' lowercase_ = fetch_california_housing() lowercase_ = data_handling(_UpperCamelCase ) lowercase_ = train_test_split( _UpperCamelCase , _UpperCamelCase , test_size=0.2_5 , random_state=1 ) lowercase_ = xgboost(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # Error printing print(F'''Mean Absolute Error : {mean_absolute_error(_UpperCamelCase , _UpperCamelCase )}''' ) print(F'''Mean Square Error : {mean_squared_error(_UpperCamelCase , _UpperCamelCase )}''' ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
30
import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('''9.1.0'''): lowerCAmelCase_ = { '''linear''': PIL.Image.Resampling.BILINEAR, '''bilinear''': PIL.Image.Resampling.BILINEAR, '''bicubic''': PIL.Image.Resampling.BICUBIC, '''lanczos''': PIL.Image.Resampling.LANCZOS, '''nearest''': PIL.Image.Resampling.NEAREST, } else: lowerCAmelCase_ = { '''linear''': PIL.Image.LINEAR, '''bilinear''': PIL.Image.BILINEAR, '''bicubic''': PIL.Image.BICUBIC, '''lanczos''': PIL.Image.LANCZOS, '''nearest''': PIL.Image.NEAREST, } def lowerCamelCase_ ( _UpperCamelCase ) -> Optional[int]: """simple docstring""" snake_case_ : Dict = (images / 2 + 0.5).clamp(0 , 1 ) snake_case_ : Dict = images.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() snake_case_ : int = numpy_to_pil(_UpperCamelCase ) return images def lowerCamelCase_ ( _UpperCamelCase ) -> List[Any]: """simple docstring""" if images.ndim == 3: snake_case_ : Optional[Any] = images[None, ...] snake_case_ : Any = (images * 255).round().astype('''uint8''' ) if images.shape[-1] == 1: # special case for grayscale (single channel) images snake_case_ : str = [Image.fromarray(image.squeeze() , mode='''L''' ) for image in images] else: snake_case_ : List[Any] = [Image.fromarray(_UpperCamelCase ) for image in images] return pil_images
279
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __A : List[str] = logging.get_logger(__name__) class _SCREAMING_SNAKE_CASE ( _a): _UpperCamelCase:Union[str, Any] = """timm_backbone""" def __init__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE , )-> List[str]: super().__init__(**_SCREAMING_SNAKE_CASE ) lowerCamelCase_ =backbone lowerCamelCase_ =num_channels lowerCamelCase_ =features_only lowerCamelCase_ =use_pretrained_backbone lowerCamelCase_ =True lowerCamelCase_ =out_indices if out_indices is not None else (-1,)
367
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": __A : Optional[int] = pd.read_csv('sample_data.csv', header=None) __A : Optional[Any] = df.shape[:1][0] # If you're using some other dataset input the target column __A : Tuple = df.iloc[:, 1:2] __A : Tuple = actual_data.values.reshape(len_data, 1) __A : str = MinMaxScaler().fit_transform(actual_data) __A : List[str] = 10 __A : Any = 5 __A : Optional[Any] = 20 __A : List[str] = len_data - periods * look_back __A : str = actual_data[:division] __A : int = actual_data[division - look_back :] __A, __A : List[str] = [], [] __A, __A : Union[str, Any] = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) __A : List[Any] = np.array(train_x) __A : Tuple = np.array(test_x) __A : Any = np.array([list(i.ravel()) for i in train_y]) __A : List[Any] = np.array([list(i.ravel()) for i in test_y]) __A : Union[str, Any] = Sequential() model.add(LSTM(1_28, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(1_28, 1))) model.add(Dense(forward_days)) model.compile(loss='mean_squared_error', optimizer='adam') __A : Tuple = model.fit( x_train, y_train, epochs=1_50, verbose=1, shuffle=True, batch_size=4 ) __A : Optional[int] = model.predict(x_test)
49
0
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionAttendAndExcitePipeline, UNetaDConditionModel, ) from diffusers.utils import load_numpy, skip_mps, slow from diffusers.utils.testing_utils import require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin SCREAMING_SNAKE_CASE__ = False @skip_mps class lowercase ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE = StableDiffusionAttendAndExcitePipeline _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = TEXT_TO_IMAGE_PARAMS _SCREAMING_SNAKE_CASE = TEXT_TO_IMAGE_BATCH_PARAMS.union({'token_indices'} ) _SCREAMING_SNAKE_CASE = TEXT_TO_IMAGE_IMAGE_PARAMS _SCREAMING_SNAKE_CASE = TEXT_TO_IMAGE_IMAGE_PARAMS @classmethod def _snake_case ( cls ) -> str: super().setUpClass() torch.use_deterministic_algorithms(lowercase ) @classmethod def _snake_case ( cls ) -> List[Any]: super().tearDownClass() torch.use_deterministic_algorithms(lowercase ) def _snake_case ( self ) -> str: torch.manual_seed(0 ) lowerCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowercase , ) lowerCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=lowercase , set_alpha_to_one=lowercase , ) torch.manual_seed(0 ) lowerCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) lowerCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) lowerCAmelCase = CLIPTextModel(lowercase ) lowerCAmelCase = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) lowerCAmelCase = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def _snake_case ( self , lowercase , lowercase=0 ) -> Optional[Any]: if str(lowercase ).startswith("""mps""" ): lowerCAmelCase = torch.manual_seed(lowercase ) else: lowerCAmelCase = torch.Generator(device=lowercase ).manual_seed(lowercase ) lowerCAmelCase = lowerCAmelCase = { """prompt""": """a cat and a frog""", """token_indices""": [2, 5], """generator""": generator, """num_inference_steps""": 1, """guidance_scale""": 6.0, """output_type""": """numpy""", """max_iter_to_alter""": 2, """thresholds""": {0: 0.7}, } return inputs def _snake_case ( self ) -> Optional[int]: lowerCAmelCase = """cpu""" lowerCAmelCase = self.get_dummy_components() lowerCAmelCase = self.pipeline_class(**lowercase ) pipe.to(lowercase ) pipe.set_progress_bar_config(disable=lowercase ) lowerCAmelCase = self.get_dummy_inputs(lowercase ) lowerCAmelCase = pipe(**lowercase ).images lowerCAmelCase = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 64, 64, 3) ) lowerCAmelCase = np.array( [0.63_905_364, 0.62_897_307, 0.48_599_017, 0.5_133_624, 0.5_550_048, 0.45_769_516, 0.50_326_973, 0.5_023_139, 0.45_384_496] ) lowerCAmelCase = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowercase , 1e-3 ) def _snake_case ( self ) -> Union[str, Any]: super().test_cpu_offload_forward_pass(expected_max_diff=5e-4 ) def _snake_case ( self ) -> int: # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def _snake_case ( self ) -> Optional[int]: self._test_inference_batch_single_identical(batch_size=2 , expected_max_diff=7e-4 ) def _snake_case ( self ) -> Optional[int]: super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def _snake_case ( self ) -> Optional[int]: super().test_pt_np_pil_outputs_equivalent(expected_max_diff=5e-4 ) def _snake_case ( self ) -> int: super().test_save_load_local(expected_max_difference=5e-4 ) def _snake_case ( self ) -> int: super().test_save_load_optional_components(expected_max_difference=4e-4 ) @require_torch_gpu @slow class lowercase ( unittest.TestCase ): @classmethod def _snake_case ( cls ) -> Dict: super().setUpClass() torch.use_deterministic_algorithms(lowercase ) @classmethod def _snake_case ( cls ) -> Tuple: super().tearDownClass() torch.use_deterministic_algorithms(lowercase ) def _snake_case ( self ) -> List[str]: super().tearDown() gc.collect() torch.cuda.empty_cache() def _snake_case ( self ) -> List[Any]: lowerCAmelCase = torch.manual_seed(51 ) lowerCAmelCase = StableDiffusionAttendAndExcitePipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , safety_checker=lowercase , torch_dtype=torch.floataa ) pipe.to("""cuda""" ) lowerCAmelCase = """a painting of an elephant with glasses""" lowerCAmelCase = [5, 7] lowerCAmelCase = pipe( prompt=lowercase , token_indices=lowercase , guidance_scale=7.5 , generator=lowercase , num_inference_steps=5 , max_iter_to_alter=5 , output_type="""numpy""" , ).images[0] lowerCAmelCase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/attend-and-excite/elephant_glasses.npy""" ) assert np.abs((expected_image - image).max() ) < 5e-1
46
'''simple docstring''' def lowercase__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )-> bool: return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(__UpperCamelCase ) ) def lowercase__ ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase )-> bool: # Base Case if index == len(__UpperCamelCase ): return True # Recursive Step for i in range(__UpperCamelCase ): if valid_coloring(graph[index] , __UpperCamelCase , __UpperCamelCase ): # Color current vertex UpperCamelCase = i # Validate coloring if util_color(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , index + 1 ): return True # Backtrack UpperCamelCase = -1 return False def lowercase__ ( __UpperCamelCase , __UpperCamelCase )-> list[int]: UpperCamelCase = [-1] * len(__UpperCamelCase ) if util_color(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , 0 ): return colored_vertices return []
321
0
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def UpperCAmelCase_( a__ ): """simple docstring""" return {key.lstrip('''-''' ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def UpperCAmelCase_( ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = ArgumentParser( '''HuggingFace Datasets CLI tool''' , usage='''datasets-cli <command> [<args>]''' , allow_abbrev=a__ ) SCREAMING_SNAKE_CASE : Tuple = parser.add_subparsers(help='''datasets-cli command helpers''' ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(a__ ) EnvironmentCommand.register_subcommand(a__ ) TestCommand.register_subcommand(a__ ) RunBeamCommand.register_subcommand(a__ ) DummyDataCommand.register_subcommand(a__ ) # Parse args SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Tuple = parser.parse_known_args() if not hasattr(a__ , '''func''' ): parser.print_help() exit(1 ) SCREAMING_SNAKE_CASE : int = parse_unknown_args(a__ ) # Run SCREAMING_SNAKE_CASE : Optional[Any] = args.func(a__ , **a__ ) service.run() if __name__ == "__main__": main()
19
def UpperCAmelCase_( a__ ): """simple docstring""" if divisor % 5 == 0 or divisor % 2 == 0: return 0 SCREAMING_SNAKE_CASE : Tuple = 1 SCREAMING_SNAKE_CASE : Tuple = 1 while repunit: SCREAMING_SNAKE_CASE : Dict = (10 * repunit + 1) % divisor repunit_index += 1 return repunit_index def UpperCAmelCase_( a__ = 1_000_000 ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = limit - 1 if divisor % 2 == 0: divisor += 1 while least_divisible_repunit(a__ ) <= limit: divisor += 2 return divisor if __name__ == "__main__": print(F"{solution() = }")
19
1
'''simple docstring''' from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig _lowercase : List[Any] = logging.get_logger(__name__) # General docstring _lowercase : Optional[Any] = "RegNetConfig" # Base docstring _lowercase : Optional[int] = "facebook/regnet-y-040" _lowercase : Union[str, Any] = [1, 1_0_8_8, 7, 7] # Image classification docstring _lowercase : Union[str, Any] = "facebook/regnet-y-040" _lowercase : int = "tabby, tabby cat" _lowercase : Union[str, Any] = [ "facebook/regnet-y-040", # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 3 , __SCREAMING_SNAKE_CASE = 1 , __SCREAMING_SNAKE_CASE = 1 , __SCREAMING_SNAKE_CASE = "relu" , ): """simple docstring""" super().__init__() lowercase_ : Dict = nn.Convad( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=__SCREAMING_SNAKE_CASE , stride=__SCREAMING_SNAKE_CASE , padding=kernel_size // 2 , groups=__SCREAMING_SNAKE_CASE , bias=__SCREAMING_SNAKE_CASE , ) lowercase_ : int = nn.BatchNormad(__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = ACTaFN[activation] if activation is not None else nn.Identity() def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Tuple = self.convolution(__SCREAMING_SNAKE_CASE ) lowercase_ : Dict = self.normalization(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = self.activation(__SCREAMING_SNAKE_CASE ) return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" super().__init__() lowercase_ : Dict = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) lowercase_ : Union[str, Any] = config.num_channels def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Union[str, Any] = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) lowercase_ : Any = self.embedder(__SCREAMING_SNAKE_CASE ) return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 2 ): """simple docstring""" super().__init__() lowercase_ : Optional[Any] = nn.Convad(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=1 , stride=__SCREAMING_SNAKE_CASE , bias=__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = nn.BatchNormad(__SCREAMING_SNAKE_CASE ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : str = self.convolution(__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = self.normalization(__SCREAMING_SNAKE_CASE ) return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" super().__init__() lowercase_ : Any = nn.AdaptiveAvgPoolad((1, 1) ) lowercase_ : Tuple = nn.Sequential( nn.Convad(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=1 ) , nn.ReLU() , nn.Convad(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=1 ) , nn.Sigmoid() , ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Optional[Any] = self.pooler(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = self.attention(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = hidden_state * attention return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 1 ): """simple docstring""" super().__init__() lowercase_ : Any = in_channels != out_channels or stride != 1 lowercase_ : str = max(1 , out_channels // config.groups_width ) lowercase_ : str = ( RegNetShortCut(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=__SCREAMING_SNAKE_CASE ) if should_apply_shortcut else nn.Identity() ) lowercase_ : int = nn.Sequential( RegNetConvLayer(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=__SCREAMING_SNAKE_CASE , groups=__SCREAMING_SNAKE_CASE , activation=config.hidden_act ) , RegNetConvLayer(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=1 , activation=__SCREAMING_SNAKE_CASE ) , ) lowercase_ : Optional[int] = ACTaFN[config.hidden_act] def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Dict = hidden_state lowercase_ : int = self.layer(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = self.shortcut(__SCREAMING_SNAKE_CASE ) hidden_state += residual lowercase_ : str = self.activation(__SCREAMING_SNAKE_CASE ) return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 1 ): """simple docstring""" super().__init__() lowercase_ : Optional[int] = in_channels != out_channels or stride != 1 lowercase_ : Dict = max(1 , out_channels // config.groups_width ) lowercase_ : Dict = ( RegNetShortCut(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=__SCREAMING_SNAKE_CASE ) if should_apply_shortcut else nn.Identity() ) lowercase_ : str = nn.Sequential( RegNetConvLayer(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=__SCREAMING_SNAKE_CASE , groups=__SCREAMING_SNAKE_CASE , activation=config.hidden_act ) , RegNetSELayer(__SCREAMING_SNAKE_CASE , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , kernel_size=1 , activation=__SCREAMING_SNAKE_CASE ) , ) lowercase_ : Optional[Any] = ACTaFN[config.hidden_act] def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Optional[int] = hidden_state lowercase_ : Optional[Any] = self.layer(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = self.shortcut(__SCREAMING_SNAKE_CASE ) hidden_state += residual lowercase_ : Union[str, Any] = self.activation(__SCREAMING_SNAKE_CASE ) return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 2 , __SCREAMING_SNAKE_CASE = 2 , ): """simple docstring""" super().__init__() lowercase_ : Any = RegNetXLayer if config.layer_type == '''x''' else RegNetYLayer lowercase_ : Any = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , stride=__SCREAMING_SNAKE_CASE , ) , *[layer(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for _ in range(depth - 1 )] , ) def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Dict = self.layers(__SCREAMING_SNAKE_CASE ) return hidden_state class lowerCAmelCase__ ( nn.Module ): def __init__( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" super().__init__() lowercase_ : Optional[Any] = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( __SCREAMING_SNAKE_CASE , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) lowercase_ : Dict = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(__SCREAMING_SNAKE_CASE , config.depths[1:] ): self.stages.append(RegNetStage(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , depth=__SCREAMING_SNAKE_CASE ) ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = False , __SCREAMING_SNAKE_CASE = True ): """simple docstring""" lowercase_ : int = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: lowercase_ : int = hidden_states + (hidden_state,) lowercase_ : List[str] = stage_module(__SCREAMING_SNAKE_CASE ) if output_hidden_states: lowercase_ : Optional[int] = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=__SCREAMING_SNAKE_CASE , hidden_states=__SCREAMING_SNAKE_CASE ) class lowerCAmelCase__ ( lowerCamelCase_ ): lowerCAmelCase_ = RegNetConfig lowerCAmelCase_ = '''regnet''' lowerCAmelCase_ = '''pixel_values''' lowerCAmelCase_ = True def _snake_case ( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" if isinstance(__SCREAMING_SNAKE_CASE , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode='''fan_out''' , nonlinearity='''relu''' ) elif isinstance(__SCREAMING_SNAKE_CASE , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=False ): """simple docstring""" if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase_ : Union[str, Any] = value _lowercase : List[Any] = r"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _lowercase : Optional[int] = r"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConvNextImageProcessor.__call__`] for details.\n\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( '''The bare RegNet model outputting raw features without any specific head on top.''' , lowerCamelCase_ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class lowerCAmelCase__ ( lowerCamelCase_ ): def __init__( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" super().__init__(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = config lowercase_ : Tuple = RegNetEmbeddings(__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = RegNetEncoder(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__SCREAMING_SNAKE_CASE ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=__SCREAMING_SNAKE_CASE , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None ): """simple docstring""" lowercase_ : Tuple = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase_ : str = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : int = self.embedder(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = self.encoder( __SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE , return_dict=__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = encoder_outputs[0] lowercase_ : List[Any] = self.pooler(__SCREAMING_SNAKE_CASE ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=__SCREAMING_SNAKE_CASE , pooler_output=__SCREAMING_SNAKE_CASE , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( ''' RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''' , lowerCamelCase_ , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class lowerCAmelCase__ ( lowerCamelCase_ ): def __init__( self , __SCREAMING_SNAKE_CASE ): """simple docstring""" super().__init__(__SCREAMING_SNAKE_CASE ) lowercase_ : str = config.num_labels lowercase_ : Optional[int] = RegNetModel(__SCREAMING_SNAKE_CASE ) # classification head lowercase_ : List[Any] = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__SCREAMING_SNAKE_CASE ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=__SCREAMING_SNAKE_CASE , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _snake_case ( self , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = None , ): """simple docstring""" lowercase_ : Optional[int] = return_dict if return_dict is not None else self.config.use_return_dict lowercase_ : str = self.regnet(__SCREAMING_SNAKE_CASE , output_hidden_states=__SCREAMING_SNAKE_CASE , return_dict=__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = outputs.pooler_output if return_dict else outputs[1] lowercase_ : Any = self.classifier(__SCREAMING_SNAKE_CASE ) lowercase_ : Any = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: lowercase_ : str = '''regression''' elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): lowercase_ : List[Any] = '''single_label_classification''' else: lowercase_ : Optional[int] = '''multi_label_classification''' if self.config.problem_type == "regression": lowercase_ : int = MSELoss() if self.num_labels == 1: lowercase_ : Dict = loss_fct(logits.squeeze() , labels.squeeze() ) else: lowercase_ : List[str] = loss_fct(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) elif self.config.problem_type == "single_label_classification": lowercase_ : List[str] = CrossEntropyLoss() lowercase_ : str = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": lowercase_ : Union[str, Any] = BCEWithLogitsLoss() lowercase_ : Optional[int] = loss_fct(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if not return_dict: lowercase_ : str = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=__SCREAMING_SNAKE_CASE , logits=__SCREAMING_SNAKE_CASE , hidden_states=outputs.hidden_states )
93
def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[str]: if index == r: for j in range(SCREAMING_SNAKE_CASE__ ): print(data[j] , end=""" """ ) print(""" """ ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location lowercase : Tuple = arr[i] combination_util(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , index + 1 , SCREAMING_SNAKE_CASE__ , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> List[str]: # A temporary array to store all combination one by one lowercase : Optional[int] = [0] * r # Print all combination using temporary array 'data[]' combination_util(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , 0 , SCREAMING_SNAKE_CASE__ , 0 ) if __name__ == "__main__": # Driver code to check the function above lowercase : int = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
20
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A = { "configuration_altclip": [ "ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "AltCLIPConfig", "AltCLIPTextConfig", "AltCLIPVisionConfig", ], "processing_altclip": ["AltCLIPProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "AltCLIPPreTrainedModel", "AltCLIPModel", "AltCLIPTextModel", "AltCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys __A = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
350
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) ->str: """simple docstring""" if version.parse(hfh.__version__ ).release < version.parse('0.11.0' ).release: # old versions of hfh don't url-encode the file path lowerCAmelCase__ :Optional[int] = quote(_SCREAMING_SNAKE_CASE ) return hfh.hf_hub_url(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type='dataset' , revision=_SCREAMING_SNAKE_CASE )
254
0
from math import sqrt def lowerCamelCase__ ( __lowerCAmelCase : int ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and ( number >= 0 ), "'number' must been an int and positive" lowerCAmelCase_ = True # 0 and 1 are none primes. if number <= 1: lowerCAmelCase_ = False for divisor in range(2 , int(round(sqrt(snake_case__ ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowerCAmelCase_ = False break # precondition assert isinstance(snake_case__ , snake_case__ ), "'status' must been from type bool" return status def lowerCamelCase__ ( __lowerCAmelCase : Tuple ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowerCAmelCase_ = list(range(2 , n + 1 ) ) lowerCAmelCase_ = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(snake_case__ ) ): for j in range(i + 1 , len(snake_case__ ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowerCAmelCase_ = 0 # filters actual prime numbers. lowerCAmelCase_ = [x for x in begin_list if x != 0] # precondition assert isinstance(snake_case__ , snake_case__ ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __lowerCAmelCase : List[str] ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and (n > 2), "'N' must been an int and > 2" lowerCAmelCase_ = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(snake_case__ ): ans.append(snake_case__ ) # precondition assert isinstance(snake_case__ , snake_case__ ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __lowerCAmelCase : Union[str, Any] ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and number >= 0, "'number' must been an int and >= 0" lowerCAmelCase_ = [] # this list will be returns of the function. # potential prime number factors. lowerCAmelCase_ = 2 lowerCAmelCase_ = number if number == 0 or number == 1: ans.append(snake_case__ ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(snake_case__ ): while quotient != 1: if is_prime(snake_case__ ) and (quotient % factor == 0): ans.append(snake_case__ ) quotient /= factor else: factor += 1 else: ans.append(snake_case__ ) # precondition assert isinstance(snake_case__ , snake_case__ ), "'ans' must been from type list" return ans def lowerCamelCase__ ( __lowerCAmelCase : Optional[Any] ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ = 0 # prime factorization of 'number' lowerCAmelCase_ = prime_factorization(snake_case__ ) lowerCAmelCase_ = max(snake_case__ ) # precondition assert isinstance(snake_case__ , snake_case__ ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __lowerCAmelCase : Union[str, Any] ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowerCAmelCase_ = 0 # prime factorization of 'number' lowerCAmelCase_ = prime_factorization(snake_case__ ) lowerCAmelCase_ = min(snake_case__ ) # precondition assert isinstance(snake_case__ , snake_case__ ), "'ans' must been from type int" return ans def lowerCamelCase__ ( __lowerCAmelCase : List[str] ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ), "'number' must been an int" assert isinstance(number % 2 == 0 , snake_case__ ), "compare bust been from type bool" return number % 2 == 0 def lowerCamelCase__ ( __lowerCAmelCase : str ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ), "'number' must been an int" assert isinstance(number % 2 != 0 , snake_case__ ), "compare bust been from type bool" return number % 2 != 0 def lowerCamelCase__ ( __lowerCAmelCase : Optional[int] ): """simple docstring""" assert ( isinstance(snake_case__ , snake_case__ ) and (number > 2) and is_even(snake_case__ ) ), "'number' must been an int, even and > 2" lowerCAmelCase_ = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowerCAmelCase_ = get_prime_numbers(snake_case__ ) lowerCAmelCase_ = len(snake_case__ ) # run variable for while-loops. lowerCAmelCase_ = 0 lowerCAmelCase_ = None # exit variable. for break up the loops lowerCAmelCase_ = True while i < len_pn and loop: lowerCAmelCase_ = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowerCAmelCase_ = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(snake_case__ , snake_case__ ) and (len(snake_case__ ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def lowerCamelCase__ ( __lowerCAmelCase : List[str] , __lowerCAmelCase : Optional[Any] ): """simple docstring""" assert ( isinstance(snake_case__ , snake_case__ ) and isinstance(snake_case__ , snake_case__ ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ = 0 while numbera != 0: lowerCAmelCase_ = numbera % numbera lowerCAmelCase_ = numbera lowerCAmelCase_ = rest # precondition assert isinstance(snake_case__ , snake_case__ ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def lowerCamelCase__ ( __lowerCAmelCase : List[Any] , __lowerCAmelCase : int ): """simple docstring""" assert ( isinstance(snake_case__ , snake_case__ ) and isinstance(snake_case__ , snake_case__ ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowerCAmelCase_ = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowerCAmelCase_ = prime_factorization(snake_case__ ) lowerCAmelCase_ = prime_factorization(snake_case__ ) elif numbera == 1 or numbera == 1: lowerCAmelCase_ = [] lowerCAmelCase_ = [] lowerCAmelCase_ = max(snake_case__ , snake_case__ ) lowerCAmelCase_ = 0 lowerCAmelCase_ = 0 lowerCAmelCase_ = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowerCAmelCase_ = prime_fac_a.count(snake_case__ ) lowerCAmelCase_ = prime_fac_a.count(snake_case__ ) for _ in range(max(snake_case__ , snake_case__ ) ): ans *= n else: lowerCAmelCase_ = prime_fac_a.count(snake_case__ ) for _ in range(snake_case__ ): ans *= n done.append(snake_case__ ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowerCAmelCase_ = prime_fac_a.count(snake_case__ ) for _ in range(snake_case__ ): ans *= n done.append(snake_case__ ) # precondition assert isinstance(snake_case__ , snake_case__ ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def lowerCamelCase__ ( __lowerCAmelCase : str ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and (n >= 0), "'number' must been a positive int" lowerCAmelCase_ = 0 lowerCAmelCase_ = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(snake_case__ ): ans += 1 # precondition assert isinstance(snake_case__ , snake_case__ ) and is_prime( snake_case__ ), "'ans' must been a prime number and from type int" return ans def lowerCamelCase__ ( __lowerCAmelCase : Any , __lowerCAmelCase : List[Any] ): """simple docstring""" assert ( is_prime(snake_case__ ) and is_prime(snake_case__ ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowerCAmelCase_ = p_number_a + 1 # jump to the next number lowerCAmelCase_ = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(snake_case__ ): number += 1 while number < p_number_a: ans.append(snake_case__ ) number += 1 # fetch the next prime number. while not is_prime(snake_case__ ): number += 1 # precondition assert ( isinstance(snake_case__ , snake_case__ ) and ans[0] != p_number_a and ans[len(snake_case__ ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def lowerCamelCase__ ( __lowerCAmelCase : int ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and (n >= 1), "'n' must been int and >= 1" lowerCAmelCase_ = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(snake_case__ ) # precondition assert ans[0] == 1 and ans[len(snake_case__ ) - 1] == n, "Error in function getDivisiors(...)" return ans def lowerCamelCase__ ( __lowerCAmelCase : Optional[Any] ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and ( number > 1 ), "'number' must been an int and >= 1" lowerCAmelCase_ = get_divisors(snake_case__ ) # precondition assert ( isinstance(snake_case__ , snake_case__ ) and (divisors[0] == 1) and (divisors[len(snake_case__ ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def lowerCamelCase__ ( __lowerCAmelCase : Tuple , __lowerCAmelCase : List[str] ): """simple docstring""" assert ( isinstance(snake_case__ , snake_case__ ) and isinstance(snake_case__ , snake_case__ ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowerCAmelCase_ = gcd(abs(snake_case__ ) , abs(snake_case__ ) ) # precondition assert ( isinstance(snake_case__ , snake_case__ ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def lowerCamelCase__ ( __lowerCAmelCase : Any ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and (n >= 0), "'n' must been a int and >= 0" lowerCAmelCase_ = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def lowerCamelCase__ ( __lowerCAmelCase : str ): """simple docstring""" assert isinstance(snake_case__ , snake_case__ ) and (n >= 0), "'n' must been an int and >= 0" lowerCAmelCase_ = 0 lowerCAmelCase_ = 1 lowerCAmelCase_ = 1 # this will be return for _ in range(n - 1 ): lowerCAmelCase_ = ans ans += fiba lowerCAmelCase_ = tmp return ans
231
from __future__ import annotations __UpperCAmelCase = { '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } class lowerCAmelCase_ : def __init__( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> None: UpperCamelCase : Union[str, Any] = graph # mapping node to its parent in resulting breadth first tree UpperCamelCase : dict[str, str | None] = {} UpperCamelCase : Union[str, Any] = source_vertex def snake_case_ ( self ) -> None: UpperCamelCase : str = {self.source_vertex} UpperCamelCase : str = None UpperCamelCase : int = [self.source_vertex] # first in first out queue while queue: UpperCamelCase : Dict = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(SCREAMING_SNAKE_CASE_ ) UpperCamelCase : Tuple = vertex queue.append(SCREAMING_SNAKE_CASE_ ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_ ) -> str: if target_vertex == self.source_vertex: return self.source_vertex UpperCamelCase : Optional[Any] = self.parent.get(SCREAMING_SNAKE_CASE_ ) if target_vertex_parent is None: UpperCamelCase : Union[str, Any] = ( F"""No path from vertex: {self.source_vertex} to vertex: {target_vertex}""" ) raise ValueError(SCREAMING_SNAKE_CASE_ ) return self.shortest_path(SCREAMING_SNAKE_CASE_ ) + F"""->{target_vertex}""" if __name__ == "__main__": __UpperCAmelCase = Graph(graph, '''G''') g.breath_first_search() print(g.shortest_path('''D''')) print(g.shortest_path('''G''')) print(g.shortest_path('''Foo'''))
119
0
import datasets from .evaluate import evaluate snake_case_ = '\\n@article{hendrycks2021cuad,\n title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review},\n author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball},\n journal={arXiv preprint arXiv:2103.06268},\n year={2021}\n}\n' snake_case_ = '\nThis metric wrap the official scoring script for version 1 of the Contract\nUnderstanding Atticus Dataset (CUAD).\nContract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510\ncommercial legal contracts that have been manually labeled to identify 41 categories of important\nclauses that lawyers look for when reviewing contracts in connection with corporate transactions.\n' snake_case_ = '\nComputes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': list of possible texts for the answer, as a list of strings\n depending on a threshold on the confidence probability of each prediction.\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the CUAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\n \'aupr\': Area Under the Precision-Recall curve\n \'prec_at_80_recall\': Precision at 80% recall\n \'prec_at_90_recall\': Precision at 90% recall\nExamples:\n >>> predictions = [{\'prediction_text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\'], \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}]\n >>> references = [{\'answers\': {\'answer_start\': [143, 49], \'text\': [\'The seller:\', \'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.\']}, \'id\': \'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties\'}]\n >>> cuad_metric = datasets.load_metric("cuad")\n >>> results = cuad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0, \'aupr\': 0.0, \'prec_at_80_recall\': 1.0, \'prec_at_90_recall\': 1.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def a (self : List[Any] ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': { '''id''': datasets.Value('''string''' ), '''prediction_text''': datasets.features.Sequence(datasets.Value('''string''' ) ), }, '''references''': { '''id''': datasets.Value('''string''' ), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string''' ), '''answer_start''': datasets.Value('''int32''' ), } ), }, } ) , codebase_urls=['''https://www.atticusprojectai.org/cuad'''] , reference_urls=['''https://www.atticusprojectai.org/cuad'''] , ) def a (self : Dict , a__ : Optional[int] , a__ : Union[str, Any] ): """simple docstring""" __snake_case = {prediction['''id''']: prediction['''prediction_text'''] for prediction in predictions} __snake_case = [ { '''paragraphs''': [ { '''qas''': [ { '''answers''': [{'''text''': answer_text} for answer_text in ref['''answers''']['''text''']], '''id''': ref['''id'''], } for ref in references ] } ] } ] __snake_case = evaluate(dataset=a__ , predictions=a__ ) return score
364
from __future__ import annotations import collections import pprint from pathlib import Path def lowerCamelCase__ ( snake_case_ : str ) -> str: return "".join(sorted(snake_case_ ) ) def lowerCamelCase__ ( snake_case_ : str ) -> list[str]: return word_by_signature[signature(snake_case_ )] snake_case_ = Path(__file__).parent.joinpath('words.txt').read_text(encoding='utf-8') snake_case_ = sorted({word.strip().lower() for word in data.splitlines()}) snake_case_ = collections.defaultdict(list) for word in word_list: word_by_signature[signature(word)].append(word) if __name__ == "__main__": snake_case_ = {word: anagram(word) for word in word_list if len(anagram(word)) > 1} with open('anagrams.txt', 'w') as file: file.write('all_anagrams = \n ') file.write(pprint.pformat(all_anagrams))
238
0
"""simple docstring""" import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowerCamelCase ( _UpperCamelCase : Union[str, Any] , _UpperCamelCase : List[str]=False ) -> Tuple: '''simple docstring''' try: __UpperCAmelCase : Dict = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCAmelCase : Union[str, Any] = default else: # KEY is set, convert it to True or False. try: __UpperCAmelCase : Union[str, Any] = strtobool(_UpperCamelCase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'''If set, {key} must be yes or no.''' ) return _value UpperCAmelCase : List[str] = parse_flag_from_env('RUN_SLOW', default=False) UpperCAmelCase : Optional[Any] = parse_flag_from_env('RUN_REMOTE', default=False) UpperCAmelCase : List[str] = parse_flag_from_env('RUN_LOCAL', default=True) UpperCAmelCase : Optional[Any] = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression UpperCAmelCase : Optional[int] = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') UpperCAmelCase : List[Any] = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') UpperCAmelCase : Optional[Any] = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio UpperCAmelCase : Optional[Any] = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam UpperCAmelCase : int = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility UpperCAmelCase : Optional[Any] = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows UpperCAmelCase : List[Any] = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def lowerCamelCase ( _UpperCamelCase : Optional[int] ) -> Optional[Any]: '''simple docstring''' try: import faiss # noqa except ImportError: __UpperCAmelCase : Dict = unittest.skip("""test requires faiss""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : Tuple ) -> Dict: '''simple docstring''' try: import regex # noqa except ImportError: __UpperCAmelCase : Union[str, Any] = unittest.skip("""test requires regex""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : int ) -> List[str]: '''simple docstring''' try: import elasticsearch # noqa except ImportError: __UpperCAmelCase : Optional[Any] = unittest.skip("""test requires elasticsearch""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : Union[str, Any] ) -> List[str]: '''simple docstring''' try: import sqlalchemy # noqa except ImportError: __UpperCAmelCase : Dict = unittest.skip("""test requires sqlalchemy""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : str ) -> Dict: '''simple docstring''' if not config.TORCH_AVAILABLE: __UpperCAmelCase : List[Any] = unittest.skip("""test requires PyTorch""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : Optional[Any] ) -> str: '''simple docstring''' if not config.TF_AVAILABLE: __UpperCAmelCase : Dict = unittest.skip("""test requires TensorFlow""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : List[Any] ) -> List[str]: '''simple docstring''' if not config.JAX_AVAILABLE: __UpperCAmelCase : Dict = unittest.skip("""test requires JAX""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : List[str] ) -> Tuple: '''simple docstring''' if not config.PIL_AVAILABLE: __UpperCAmelCase : Optional[int] = unittest.skip("""test requires Pillow""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : Optional[int] ) -> int: '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip("""test requires transformers""" )(_UpperCamelCase ) else: return test_case def lowerCamelCase ( _UpperCamelCase : List[Any] ) -> Optional[int]: '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip("""test requires tiktoken""" )(_UpperCamelCase ) else: return test_case def lowerCamelCase ( _UpperCamelCase : List[Any] ) -> int: '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip("""test requires spacy""" )(_UpperCamelCase ) else: return test_case def lowerCamelCase ( _UpperCamelCase : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' def _require_spacy_model(_UpperCamelCase : str ): try: import spacy # noqa F401 spacy.load(_UpperCamelCase ) except ImportError: return unittest.skip("""test requires spacy""" )(_UpperCamelCase ) except OSError: return unittest.skip("""test requires spacy model '{}'""".format(_UpperCamelCase ) )(_UpperCamelCase ) else: return test_case return _require_spacy_model def lowerCamelCase ( _UpperCamelCase : str ) -> Dict: '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip("""test requires pyspark""" )(_UpperCamelCase ) else: return test_case def lowerCamelCase ( _UpperCamelCase : Any ) -> Dict: '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip("""test requires joblibspark""" )(_UpperCamelCase ) else: return test_case def lowerCamelCase ( _UpperCamelCase : str ) -> Union[str, Any]: '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: __UpperCAmelCase : List[Any] = unittest.skip("""test is slow""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : Optional[int] ) -> Optional[int]: '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: __UpperCAmelCase : List[Any] = unittest.skip("""test is local""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : List[str] ) -> List[Any]: '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCAmelCase : int = unittest.skip("""test is packaged""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( _UpperCamelCase : Dict ) -> Dict: '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: __UpperCAmelCase : Optional[Any] = unittest.skip("""test requires remote""" )(_UpperCamelCase ) return test_case def lowerCamelCase ( *_UpperCamelCase : Union[str, Any] ) -> str: '''simple docstring''' def decorate(cls : Dict ): for name, fn in cls.__dict__.items(): if callable(_UpperCamelCase ) and name.startswith("""test""" ): for decorator in decorators: __UpperCAmelCase : Any = decorator(_UpperCamelCase ) setattr(cls , _UpperCamelCase , _UpperCamelCase ) return cls return decorate class lowerCamelCase__ ( A ): """simple docstring""" pass class lowerCamelCase__ ( A ): """simple docstring""" __a = 0 __a = 1 __a = 2 @contextmanager def lowerCamelCase ( _UpperCamelCase : List[str]=OfflineSimulationMode.CONNECTION_FAILS , _UpperCamelCase : List[Any]=1E-16 ) -> int: '''simple docstring''' __UpperCAmelCase : int = requests.Session().request def timeout_request(_UpperCamelCase : Tuple , _UpperCamelCase : int , _UpperCamelCase : List[str] , **_UpperCamelCase : Tuple ): # Change the url to an invalid url so that the connection hangs __UpperCAmelCase : int = """https://10.255.255.1""" if kwargs.get("""timeout""" ) is None: raise RequestWouldHangIndefinitelyError( f'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCAmelCase : List[Any] = timeout try: return online_request(_UpperCamelCase , _UpperCamelCase , **_UpperCamelCase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCAmelCase : Optional[int] = url __UpperCAmelCase : Union[str, Any] = e.args[0] __UpperCAmelCase : Union[str, Any] = (max_retry_error.args[0].replace("""10.255.255.1""" , f'''OfflineMock[{url}]''' ),) __UpperCAmelCase : Any = (max_retry_error,) raise def raise_connection_error(_UpperCamelCase : str , _UpperCamelCase : int , **_UpperCamelCase : Optional[int] ): raise requests.ConnectionError("""Offline mode is enabled.""" , request=_UpperCamelCase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("""requests.Session.send""" , _UpperCamelCase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("""requests.Session.request""" , _UpperCamelCase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("""datasets.config.HF_DATASETS_OFFLINE""" , _UpperCamelCase ): yield else: raise ValueError("""Please use a value from the OfflineSimulationMode enum.""" ) @contextmanager def lowerCamelCase ( *_UpperCamelCase : int , **_UpperCamelCase : Any ) -> Dict: '''simple docstring''' __UpperCAmelCase : Any = str(Path().resolve() ) with tempfile.TemporaryDirectory(*_UpperCamelCase , **_UpperCamelCase ) as tmp_dir: try: os.chdir(_UpperCamelCase ) yield finally: os.chdir(_UpperCamelCase ) @contextmanager def lowerCamelCase ( ) -> Union[str, Any]: '''simple docstring''' import gc gc.collect() __UpperCAmelCase : Union[str, Any] = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowerCamelCase ( ) -> str: '''simple docstring''' import gc gc.collect() __UpperCAmelCase : Optional[Any] = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowerCamelCase ( _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Union[str, Any] ) -> Tuple: '''simple docstring''' return deepcopy(_UpperCamelCase ).integers(0 , 1_0_0 , 1_0 ).tolist() == deepcopy(_UpperCamelCase ).integers(0 , 1_0_0 , 1_0 ).tolist() def lowerCamelCase ( _UpperCamelCase : List[Any] ) -> Optional[Any]: '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(_UpperCamelCase : Dict , *_UpperCamelCase : Optional[Any] , **_UpperCamelCase : Any ): try: return func(*_UpperCamelCase , **_UpperCamelCase ) except HTTPError as err: if str(_UpperCamelCase ).startswith("""500""" ) or str(_UpperCamelCase ).startswith("""502""" ): pytest.xfail(str(_UpperCamelCase ) ) raise err return decorator.decorator(_wrapper , _UpperCamelCase ) class lowerCamelCase__ : """simple docstring""" def __init__( self : List[Any] , UpperCamelCase : Optional[Any] , UpperCamelCase : Optional[int] , UpperCamelCase : Optional[Any] ): '''simple docstring''' __UpperCAmelCase : Optional[int] = returncode __UpperCAmelCase : Union[str, Any] = stdout __UpperCAmelCase : Any = stderr async def lowerCamelCase ( _UpperCamelCase : Tuple , _UpperCamelCase : Optional[int] ) -> List[Any]: '''simple docstring''' while True: __UpperCAmelCase : str = await stream.readline() if line: callback(_UpperCamelCase ) else: break async def lowerCamelCase ( _UpperCamelCase : Optional[int] , _UpperCamelCase : int=None , _UpperCamelCase : Optional[Any]=None , _UpperCamelCase : List[Any]=None , _UpperCamelCase : str=False , _UpperCamelCase : List[Any]=False ) -> _RunOutput: '''simple docstring''' if echo: print("""\nRunning: """ , """ """.join(_UpperCamelCase ) ) __UpperCAmelCase : Tuple = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=_UpperCamelCase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=_UpperCamelCase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCAmelCase : Any = [] __UpperCAmelCase : Optional[int] = [] def tee(_UpperCamelCase : Optional[Any] , _UpperCamelCase : List[str] , _UpperCamelCase : List[str] , _UpperCamelCase : int="" ): __UpperCAmelCase : Union[str, Any] = line.decode("""utf-8""" ).rstrip() sink.append(_UpperCamelCase ) if not quiet: print(_UpperCamelCase , _UpperCamelCase , file=_UpperCamelCase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda _UpperCamelCase : tee(_UpperCamelCase , _UpperCamelCase , sys.stdout , label="""stdout:""" ) ), _read_stream(p.stderr , lambda _UpperCamelCase : tee(_UpperCamelCase , _UpperCamelCase , sys.stderr , label="""stderr:""" ) ), ] , timeout=_UpperCamelCase , ) return _RunOutput(await p.wait() , _UpperCamelCase , _UpperCamelCase ) def lowerCamelCase ( _UpperCamelCase : str , _UpperCamelCase : Optional[Any]=None , _UpperCamelCase : Dict=None , _UpperCamelCase : List[Any]=1_8_0 , _UpperCamelCase : Dict=False , _UpperCamelCase : List[Any]=True ) -> _RunOutput: '''simple docstring''' __UpperCAmelCase : Optional[int] = asyncio.get_event_loop() __UpperCAmelCase : Union[str, Any] = loop.run_until_complete( _stream_subprocess(_UpperCamelCase , env=_UpperCamelCase , stdin=_UpperCamelCase , timeout=_UpperCamelCase , quiet=_UpperCamelCase , echo=_UpperCamelCase ) ) __UpperCAmelCase : int = """ """.join(_UpperCamelCase ) if result.returncode > 0: __UpperCAmelCase : Union[str, Any] = """\n""".join(result.stderr ) raise RuntimeError( f'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' f'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f'''\'{cmd_str}\' produced no output.''' ) return result def lowerCamelCase ( ) -> str: '''simple docstring''' __UpperCAmelCase : Dict = os.environ.get("""PYTEST_XDIST_WORKER""" , """gw0""" ) __UpperCAmelCase : Dict = re.sub(R"""^gw""" , """""" , _UpperCamelCase , 0 , re.M ) return int(_UpperCamelCase ) def lowerCamelCase ( ) -> List[str]: '''simple docstring''' __UpperCAmelCase : Optional[Any] = 2_9_5_0_0 __UpperCAmelCase : int = pytest_xdist_worker_id() return port + uniq_delta
115
"""simple docstring""" import math def lowerCamelCase ( _UpperCamelCase : int ) -> list[int]: '''simple docstring''' __UpperCAmelCase : List[Any] = [] __UpperCAmelCase : Dict = 2 __UpperCAmelCase : Union[str, Any] = int(math.sqrt(_UpperCamelCase ) ) # Size of every segment __UpperCAmelCase : Tuple = [True] * (end + 1) __UpperCAmelCase : int = [] while start <= end: if temp[start] is True: in_prime.append(_UpperCamelCase ) for i in range(start * start , end + 1 , _UpperCamelCase ): __UpperCAmelCase : Dict = False start += 1 prime += in_prime __UpperCAmelCase : Optional[int] = end + 1 __UpperCAmelCase : Dict = min(2 * end , _UpperCamelCase ) while low <= n: __UpperCAmelCase : Union[str, Any] = [True] * (high - low + 1) for each in in_prime: __UpperCAmelCase : Dict = math.floor(low / each ) * each if t < low: t += each for j in range(_UpperCamelCase , high + 1 , _UpperCamelCase ): __UpperCAmelCase : Tuple = False for j in range(len(_UpperCamelCase ) ): if temp[j] is True: prime.append(j + low ) __UpperCAmelCase : Tuple = high + 1 __UpperCAmelCase : Optional[int] = min(high + end , _UpperCamelCase ) return prime print(sieve(10**6))
115
1
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer __magic_name__ = logging.get_logger(__name__) class lowercase ( A__ ): '''simple docstring''' __SCREAMING_SNAKE_CASE = """AutoTokenizer""" __SCREAMING_SNAKE_CASE = ["""tokenizer"""] __SCREAMING_SNAKE_CASE = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self , _snake_case , _snake_case=None ) -> Tuple: """simple docstring""" super().__init__(_snake_case ) UpperCAmelCase = speaker_embeddings @classmethod def snake_case_ ( cls , _snake_case , _snake_case="speaker_embeddings_path.json" , **_snake_case ) -> Optional[int]: """simple docstring""" if speaker_embeddings_dict_path is not None: UpperCAmelCase = get_file_from_repo( _snake_case , _snake_case , subfolder=kwargs.pop('''subfolder''' , _snake_case ) , cache_dir=kwargs.pop('''cache_dir''' , _snake_case ) , force_download=kwargs.pop('''force_download''' , _snake_case ) , proxies=kwargs.pop('''proxies''' , _snake_case ) , resume_download=kwargs.pop('''resume_download''' , _snake_case ) , local_files_only=kwargs.pop('''local_files_only''' , _snake_case ) , use_auth_token=kwargs.pop('''use_auth_token''' , _snake_case ) , revision=kwargs.pop('''revision''' , _snake_case ) , ) if speaker_embeddings_path is None: logger.warning( f"""`{os.path.join(_snake_case , _snake_case )}` does not exists , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.""" ) UpperCAmelCase = None else: with open(_snake_case ) as speaker_embeddings_json: UpperCAmelCase = json.load(_snake_case ) else: UpperCAmelCase = None UpperCAmelCase = AutoTokenizer.from_pretrained(_snake_case , **_snake_case ) return cls(tokenizer=_snake_case , speaker_embeddings=_snake_case ) def snake_case_ ( self , _snake_case , _snake_case="speaker_embeddings_path.json" , _snake_case="speaker_embeddings" , _snake_case = False , **_snake_case , ) -> Dict: """simple docstring""" if self.speaker_embeddings is not None: os.makedirs(os.path.join(_snake_case , _snake_case , '''v2''' ) , exist_ok=_snake_case ) UpperCAmelCase = {} UpperCAmelCase = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": UpperCAmelCase = self._load_voice_preset(_snake_case ) UpperCAmelCase = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict['''repo_or_path'''] , _snake_case , f"""{prompt_key}_{key}""" ) , voice_preset[key] , allow_pickle=_snake_case , ) UpperCAmelCase = os.path.join(_snake_case , f"""{prompt_key}_{key}.npy""" ) UpperCAmelCase = tmp_dict with open(os.path.join(_snake_case , _snake_case ) , '''w''' ) as fp: json.dump(_snake_case , _snake_case ) super().save_pretrained(_snake_case , _snake_case , **_snake_case ) def snake_case_ ( self , _snake_case = None , **_snake_case ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.speaker_embeddings[voice_preset] UpperCAmelCase = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( f"""Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].""" ) UpperCAmelCase = get_file_from_repo( self.speaker_embeddings.get('''repo_or_path''' , '''/''' ) , voice_preset_paths[key] , subfolder=kwargs.pop('''subfolder''' , _snake_case ) , cache_dir=kwargs.pop('''cache_dir''' , _snake_case ) , force_download=kwargs.pop('''force_download''' , _snake_case ) , proxies=kwargs.pop('''proxies''' , _snake_case ) , resume_download=kwargs.pop('''resume_download''' , _snake_case ) , local_files_only=kwargs.pop('''local_files_only''' , _snake_case ) , use_auth_token=kwargs.pop('''use_auth_token''' , _snake_case ) , revision=kwargs.pop('''revision''' , _snake_case ) , ) if path is None: raise ValueError( f"""`{os.path.join(self.speaker_embeddings.get('repo_or_path' , '/' ) , voice_preset_paths[key] )}` does not exists , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset} embeddings.""" ) UpperCAmelCase = np.load(_snake_case ) return voice_preset_dict def snake_case_ ( self , _snake_case = None ) -> Tuple: """simple docstring""" for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(f"""Voice preset unrecognized, missing {key} as a key.""" ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(f"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(f"""{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.""" ) def __call__( self , _snake_case=None , _snake_case=None , _snake_case="pt" , _snake_case=256 , _snake_case=False , _snake_case=True , _snake_case=False , **_snake_case , ) -> int: """simple docstring""" if voice_preset is not None and not isinstance(_snake_case , _snake_case ): if ( isinstance(_snake_case , _snake_case ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): UpperCAmelCase = self._load_voice_preset(_snake_case ) else: if isinstance(_snake_case , _snake_case ) and not voice_preset.endswith('''.npz''' ): UpperCAmelCase = voice_preset + '''.npz''' UpperCAmelCase = np.load(_snake_case ) if voice_preset is not None: self._validate_voice_preset_dict(_snake_case , **_snake_case ) UpperCAmelCase = BatchFeature(data=_snake_case , tensor_type=_snake_case ) UpperCAmelCase = self.tokenizer( _snake_case , return_tensors=_snake_case , padding='''max_length''' , max_length=_snake_case , return_attention_mask=_snake_case , return_token_type_ids=_snake_case , add_special_tokens=_snake_case , **_snake_case , ) if voice_preset is not None: UpperCAmelCase = voice_preset return encoded_text
152
import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __magic_name__ = get_tests_dir("fixtures/test_sentencepiece_no_bos.model") @require_sentencepiece @require_tokenizers class lowercase ( A__ , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = PegasusTokenizer __SCREAMING_SNAKE_CASE = PegasusTokenizerFast __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = True def snake_case_ ( self ) -> Optional[Any]: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = PegasusTokenizer(_snake_case ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def snake_case_ ( self ) -> Dict: """simple docstring""" return PegasusTokenizer.from_pretrained('''google/pegasus-large''' ) def snake_case_ ( self , **_snake_case ) -> PegasusTokenizer: """simple docstring""" return PegasusTokenizer.from_pretrained(self.tmpdirname , **_snake_case ) def snake_case_ ( self , _snake_case ) -> Any: """simple docstring""" return ("This is a test", "This is a test") def snake_case_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = '''</s>''' UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_snake_case ) , _snake_case ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_snake_case ) , _snake_case ) def snake_case_ ( self ) -> str: """simple docstring""" UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''</s>''' ) self.assertEqual(vocab_keys[-1] , '''v''' ) self.assertEqual(len(_snake_case ) , 1103 ) def snake_case_ ( self ) -> Union[str, Any]: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1103 ) def snake_case_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = self.tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = ( '''Let\'s see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important''' ''' </s> <pad> <pad> <pad>''' ) UpperCAmelCase = rust_tokenizer([raw_input_str] , return_tensors=_snake_case , add_special_tokens=_snake_case ).input_ids[0] UpperCAmelCase = py_tokenizer([raw_input_str] , return_tensors=_snake_case , add_special_tokens=_snake_case ).input_ids[0] self.assertListEqual(_snake_case , _snake_case ) def snake_case_ ( self ) -> Tuple: """simple docstring""" UpperCAmelCase = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word UpperCAmelCase = '''<mask_1> To ensure a <mask_2> flow of bank resolutions.''' UpperCAmelCase = [2, 413, 615, 114, 3, 1971, 113, 1679, 1_0710, 107, 1] UpperCAmelCase = tokenizer([raw_input_str] , return_tensors=_snake_case ).input_ids[0] self.assertListEqual(_snake_case , _snake_case ) def snake_case_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_6103 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 103 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 105 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1024 UpperCAmelCase = '''To ensure a smooth flow of bank resolutions.''' UpperCAmelCase = [413, 615, 114, 2291, 1971, 113, 1679, 1_0710, 107, 1] UpperCAmelCase = tokenizer([raw_input_str] , return_tensors=_snake_case ).input_ids[0] self.assertListEqual(_snake_case , _snake_case ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def snake_case_ ( self ) -> List[str]: """simple docstring""" UpperCAmelCase = ['''This is going to be way too long.''' * 150, '''short example'''] UpperCAmelCase = ['''not super long but more than 5 tokens''', '''tiny'''] UpperCAmelCase = self._large_tokenizer(_snake_case , padding=_snake_case , truncation=_snake_case , return_tensors='''pt''' ) UpperCAmelCase = self._large_tokenizer( text_target=_snake_case , max_length=5 , padding=_snake_case , truncation=_snake_case , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 1024) assert batch.attention_mask.shape == (2, 1024) assert targets["input_ids"].shape == (2, 5) assert len(_snake_case ) == 2 # input_ids, attention_mask. @slow def snake_case_ ( self ) -> List[str]: """simple docstring""" # fmt: off UpperCAmelCase = {'''input_ids''': [[3_8979, 143, 1_8485, 606, 130, 2_6669, 8_7686, 121, 5_4189, 1129, 111, 2_6669, 8_7686, 121, 9114, 1_4787, 121, 1_3249, 158, 592, 956, 121, 1_4621, 3_1576, 143, 6_2613, 108, 9688, 930, 4_3430, 1_1562, 6_2613, 304, 108, 1_1443, 897, 108, 9314, 1_7415, 6_3399, 108, 1_1443, 7614, 1_8316, 118, 4284, 7148, 1_2430, 143, 1400, 2_5703, 158, 111, 4284, 7148, 1_1772, 143, 2_1297, 1064, 158, 122, 204, 3506, 1754, 1133, 1_4787, 1581, 115, 3_3224, 4482, 111, 1355, 110, 2_9173, 317, 5_0833, 108, 2_0147, 9_4665, 111, 7_7198, 107, 1], [110, 6_2613, 117, 638, 112, 1133, 121, 2_0098, 1355, 7_9050, 1_3872, 135, 1596, 5_3541, 1352, 141, 1_3039, 5542, 124, 302, 518, 111, 268, 2956, 115, 149, 4427, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [139, 1235, 2799, 1_8289, 1_7780, 204, 109, 9474, 1296, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_snake_case , model_name='''google/bigbird-pegasus-large-arxiv''' , revision='''ba85d0851d708441f91440d509690f1ab6353415''' , ) @require_sentencepiece @require_tokenizers class lowercase ( A__ , unittest.TestCase ): '''simple docstring''' __SCREAMING_SNAKE_CASE = PegasusTokenizer __SCREAMING_SNAKE_CASE = PegasusTokenizerFast __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = True def snake_case_ ( self ) -> int: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = PegasusTokenizer(_snake_case , offset=0 , mask_token_sent=_snake_case , mask_token='''[MASK]''' ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def snake_case_ ( self ) -> Dict: """simple docstring""" return PegasusTokenizer.from_pretrained('''google/bigbird-pegasus-large-arxiv''' ) def snake_case_ ( self , **_snake_case ) -> PegasusTokenizer: """simple docstring""" return PegasusTokenizer.from_pretrained(self.tmpdirname , **_snake_case ) def snake_case_ ( self , _snake_case ) -> List[str]: """simple docstring""" return ("This is a test", "This is a test") def snake_case_ ( self ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = self.tokenizer_class.from_pretrained(self.tmpdirname ) UpperCAmelCase = ( '''Let\'s see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>''' ''' <pad> <pad> <pad>''' ) UpperCAmelCase = rust_tokenizer([raw_input_str] , return_tensors=_snake_case , add_special_tokens=_snake_case ).input_ids[0] UpperCAmelCase = py_tokenizer([raw_input_str] , return_tensors=_snake_case , add_special_tokens=_snake_case ).input_ids[0] self.assertListEqual(_snake_case , _snake_case ) @require_torch def snake_case_ ( self ) -> Optional[Any]: """simple docstring""" UpperCAmelCase = ['''This is going to be way too long.''' * 1000, '''short example'''] UpperCAmelCase = ['''not super long but more than 5 tokens''', '''tiny'''] UpperCAmelCase = self._large_tokenizer(_snake_case , padding=_snake_case , truncation=_snake_case , return_tensors='''pt''' ) UpperCAmelCase = self._large_tokenizer( text_target=_snake_case , max_length=5 , padding=_snake_case , truncation=_snake_case , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 4096) assert batch.attention_mask.shape == (2, 4096) assert targets["input_ids"].shape == (2, 5) assert len(_snake_case ) == 2 # input_ids, attention_mask. def snake_case_ ( self ) -> Any: """simple docstring""" UpperCAmelCase = ( '''This is an example string that is used to test the original TF implementation against the HF''' ''' implementation''' ) UpperCAmelCase = self._large_tokenizer(_snake_case ).input_ids self.assertListEqual( _snake_case , [182, 117, 142, 587, 4211, 120, 117, 263, 112, 804, 109, 856, 2_5016, 3137, 464, 109, 2_6955, 3137, 1] , )
152
1
'''simple docstring''' from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
28
"""simple docstring""" from collections import defaultdict class _UpperCAmelCase: def __init__( self , __a , __a) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = total # total no of tasks (N) # DP table will have a dimension of (2^M)*N # initially all values are set to -1 _UpperCamelCase = [ [-1 for i in range(total + 1)] for j in range(2 ** len(__a)) ] _UpperCamelCase = defaultdict(__a) # stores the list of persons for each task # final_mask is used to check if all persons are included by setting all bits # to 1 _UpperCamelCase = (1 << len(__a)) - 1 def UpperCAmelCase ( self , __a , __a) -> Dict: '''simple docstring''' # if mask == self.finalmask all persons are distributed tasks, return 1 if mask == self.final_mask: return 1 # if not everyone gets the task and no more tasks are available, return 0 if task_no > self.total_tasks: return 0 # if case already considered if self.dp[mask][task_no] != -1: return self.dp[mask][task_no] # Number of ways when we don't this task in the arrangement _UpperCamelCase = self.count_ways_until(__a , task_no + 1) # now assign the tasks one by one to all possible persons and recursively # assign for the remaining tasks. if task_no in self.task: for p in self.task[task_no]: # if p is already given a task if mask & (1 << p): continue # assign this task to p and change the mask value. And recursively # assign tasks with the new mask value. total_ways_util += self.count_ways_until(mask | (1 << p) , task_no + 1) # save the value. _UpperCamelCase = total_ways_util return self.dp[mask][task_no] def UpperCAmelCase ( self , __a) -> int: '''simple docstring''' # Store the list of persons for each task for i in range(len(__a)): for j in task_performed[i]: self.task[j].append(__a) # call the function to fill the DP table, final answer is stored in dp[0][1] return self.count_ways_until(0 , 1) if __name__ == "__main__": _a = 5 # total no of tasks (the value of N) # the list of tasks that can be done by M persons. _a = [[1, 3, 4], [1, 2, 5], [3, 4]] print( AssignmentUsingBitmask(task_performed, total_tasks).count_no_of_ways( task_performed ) )
194
0
"""simple docstring""" import argparse import os from pathlib import Path import torch from bark.generation import _load_model as _bark_load_model from huggingface_hub import hf_hub_download from transformers import EncodecConfig, EncodecModel, set_seed from transformers.models.bark.configuration_bark import ( BarkCoarseConfig, BarkConfig, BarkFineConfig, BarkSemanticConfig, ) from transformers.models.bark.generation_configuration_bark import ( BarkCoarseGenerationConfig, BarkFineGenerationConfig, BarkGenerationConfig, BarkSemanticGenerationConfig, ) from transformers.models.bark.modeling_bark import BarkCoarseModel, BarkFineModel, BarkModel, BarkSemanticModel from transformers.utils import logging logging.set_verbosity_info() _snake_case = logging.get_logger(__name__) set_seed(770) _snake_case = { 'c_attn': 'att_proj', 'c_proj': 'out_proj', 'c_fc': 'in_proj', 'transformer.': '', 'h.': 'layers.', 'ln_1': 'layernorm_1', 'ln_2': 'layernorm_2', 'ln_f': 'layernorm_final', 'wpe': 'position_embeds_layer', 'wte': 'input_embeds_layer', } _snake_case = { 'text_small': { 'repo_id': 'suno/bark', 'file_name': 'text.pt', }, 'coarse_small': { 'repo_id': 'suno/bark', 'file_name': 'coarse.pt', }, 'fine_small': { 'repo_id': 'suno/bark', 'file_name': 'fine.pt', }, 'text': { 'repo_id': 'suno/bark', 'file_name': 'text_2.pt', }, 'coarse': { 'repo_id': 'suno/bark', 'file_name': 'coarse_2.pt', }, 'fine': { 'repo_id': 'suno/bark', 'file_name': 'fine_2.pt', }, } _snake_case = os.path.dirname(os.path.abspath(__file__)) _snake_case = os.path.join(os.path.expanduser('~'), '.cache') _snake_case = os.path.join(os.getenv('XDG_CACHE_HOME', default_cache_dir), 'suno', 'bark_v0') def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=False ): '''simple docstring''' _a : Tuple = model_type if use_small: key += "_small" return os.path.join(UpperCamelCase__ , REMOTE_MODEL_PATHS[key]["""file_name"""] ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) hf_hub_download(repo_id=UpperCamelCase__ , filename=UpperCamelCase__ , local_dir=UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__="text" ): '''simple docstring''' if model_type == "text": _a : Optional[Any] = BarkSemanticModel _a : Union[str, Any] = BarkSemanticConfig _a : Dict = BarkSemanticGenerationConfig elif model_type == "coarse": _a : int = BarkCoarseModel _a : Union[str, Any] = BarkCoarseConfig _a : str = BarkCoarseGenerationConfig elif model_type == "fine": _a : Dict = BarkFineModel _a : List[Any] = BarkFineConfig _a : str = BarkFineGenerationConfig else: raise NotImplementedError() _a : int = F"""{model_type}_small""" if use_small else model_type _a : Optional[int] = REMOTE_MODEL_PATHS[model_key] if not os.path.exists(UpperCamelCase__ ): logger.info(F"""{model_type} model not found, downloading into `{CACHE_DIR}`.""" ) _download(model_info["""repo_id"""] , model_info["""file_name"""] ) _a : List[Any] = torch.load(UpperCamelCase__ , map_location=UpperCamelCase__ ) # this is a hack _a : Union[str, Any] = checkpoint["""model_args"""] if "input_vocab_size" not in model_args: _a : Any = model_args["""vocab_size"""] _a : Optional[int] = model_args["""vocab_size"""] del model_args["vocab_size"] # convert Bark model arguments to HF Bark model arguments _a : Dict = model_args.pop("""n_head""" ) _a : str = model_args.pop("""n_embd""" ) _a : int = model_args.pop("""n_layer""" ) _a : Any = ConfigClass(**checkpoint["""model_args"""] ) _a : Optional[Any] = ModelClass(config=UpperCamelCase__ ) _a : List[str] = GenerationConfigClass() _a : Union[str, Any] = model_generation_config _a : Tuple = checkpoint["""model"""] # fixup checkpoint _a : str = """_orig_mod.""" for k, v in list(state_dict.items() ): if k.startswith(UpperCamelCase__ ): # replace part of the key with corresponding layer name in HF implementation _a : List[Any] = k[len(UpperCamelCase__ ) :] for old_layer_name in new_layer_name_dict: _a : Dict = new_k.replace(UpperCamelCase__ , new_layer_name_dict[old_layer_name] ) _a : Optional[Any] = state_dict.pop(UpperCamelCase__ ) _a : Dict = set(state_dict.keys() ) - set(model.state_dict().keys() ) _a : Tuple = {k for k in extra_keys if not k.endswith(""".attn.bias""" )} _a : List[Any] = set(model.state_dict().keys() ) - set(state_dict.keys() ) _a : int = {k for k in missing_keys if not k.endswith(""".attn.bias""" )} if len(UpperCamelCase__ ) != 0: raise ValueError(F"""extra keys found: {extra_keys}""" ) if len(UpperCamelCase__ ) != 0: raise ValueError(F"""missing keys: {missing_keys}""" ) model.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) _a : int = model.num_parameters(exclude_embeddings=UpperCamelCase__ ) _a : Union[str, Any] = checkpoint["""best_val_loss"""].item() logger.info(F"""model loaded: {round(n_params/1e6 , 1 )}M params, {round(UpperCamelCase__ , 3 )} loss""" ) model.eval() model.to(UpperCamelCase__ ) del checkpoint, state_dict return model def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__="text" ): '''simple docstring''' if model_type not in ("text", "coarse", "fine"): raise NotImplementedError() _a : Any = """cpu""" # do conversion on cpu _a : Union[str, Any] = _get_ckpt_path(UpperCamelCase__ , use_small=UpperCamelCase__ ) _a : Any = _load_model(UpperCamelCase__ , UpperCamelCase__ , model_type=UpperCamelCase__ , use_small=UpperCamelCase__ ) # load bark initial model _a : List[str] = _bark_load_model(UpperCamelCase__ , """cpu""" , model_type=UpperCamelCase__ , use_small=UpperCamelCase__ ) if model_type == "text": _a : Any = bark_model["""model"""] if model.num_parameters(exclude_embeddings=UpperCamelCase__ ) != bark_model.get_num_params(): raise ValueError("""initial and new models don't have the same number of parameters""" ) # check if same output as the bark model _a : Dict = 5 _a : List[str] = 1_0 if model_type in ["text", "coarse"]: _a : List[Any] = torch.randint(2_5_6 , (batch_size, sequence_length) , dtype=torch.int ) _a : Dict = bark_model(UpperCamelCase__ )[0] _a : Any = model(UpperCamelCase__ ) # take last logits _a : List[Any] = output_new_model_total.logits[:, [-1], :] else: _a : Optional[Any] = 3 _a : int = 8 _a : List[str] = torch.randint(2_5_6 , (batch_size, sequence_length, n_codes_total) , dtype=torch.int ) _a : Union[str, Any] = model(UpperCamelCase__ , UpperCamelCase__ ) _a : Any = bark_model(UpperCamelCase__ , UpperCamelCase__ ) _a : List[Any] = output_new_model_total.logits # output difference should come from the difference of self-attention implementation design if output_new_model.shape != output_old_model.shape: raise ValueError("""initial and new outputs don't have the same shape""" ) if (output_new_model - output_old_model).abs().max().item() > 1e-3: raise ValueError("""initial and new outputs are not equal""" ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) def lowerCAmelCase__ ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , ): '''simple docstring''' _a : List[str] = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) _a : List[Any] = BarkSemanticConfig.from_pretrained(os.path.join(UpperCamelCase__ , """config.json""" ) ) _a : Optional[int] = BarkCoarseConfig.from_pretrained(os.path.join(UpperCamelCase__ , """config.json""" ) ) _a : int = BarkFineConfig.from_pretrained(os.path.join(UpperCamelCase__ , """config.json""" ) ) _a : Optional[Any] = EncodecConfig.from_pretrained("""facebook/encodec_24khz""" ) _a : Any = BarkSemanticModel.from_pretrained(UpperCamelCase__ ) _a : Any = BarkCoarseModel.from_pretrained(UpperCamelCase__ ) _a : List[str] = BarkFineModel.from_pretrained(UpperCamelCase__ ) _a : List[Any] = EncodecModel.from_pretrained("""facebook/encodec_24khz""" ) _a : Union[str, Any] = BarkConfig.from_sub_model_configs( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) _a : List[Any] = BarkGenerationConfig.from_sub_model_configs( semantic.generation_config , coarseAcoustic.generation_config , fineAcoustic.generation_config ) _a : List[str] = BarkModel(UpperCamelCase__ ) _a : int = semantic _a : Dict = coarseAcoustic _a : Union[str, Any] = fineAcoustic _a : str = codec _a : Optional[Any] = bark_generation_config Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) bark.save_pretrained(UpperCamelCase__ , repo_id=UpperCamelCase__ , push_to_hub=UpperCamelCase__ ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument('model_type', type=str, help='text, coarse or fine.') parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--is_small', action='store_true', help='convert the small version instead of the large.') _snake_case = parser.parse_args() load_model(args.pytorch_dump_folder_path, model_type=args.model_type, use_small=args.is_small)
324
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _snake_case = { 'configuration_vision_encoder_decoder': ['VisionEncoderDecoderConfig', 'VisionEncoderDecoderOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['VisionEncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['TFVisionEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = ['FlaxVisionEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys _snake_case = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
324
1
'''simple docstring''' import string def UpperCamelCase_( snake_case : str ): '''simple docstring''' snake_case_ = "" for i in sequence: snake_case_ = ord(snake_case ) if 6_5 <= extract <= 9_0: output += chr(1_5_5 - extract ) elif 9_7 <= extract <= 1_2_2: output += chr(2_1_9 - extract ) else: output += i return output def UpperCamelCase_( snake_case : str ): '''simple docstring''' snake_case_ = string.ascii_letters snake_case_ = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(snake_case )] if c in letters else c for c in sequence ) def UpperCamelCase_( ): '''simple docstring''' from timeit import timeit print("Running performance benchmarks..." ) snake_case_ = "from string import printable ; from __main__ import atbash, atbash_slow" print(f'> atbash_slow(): {timeit("atbash_slow(printable)" , setup=snake_case )} seconds' ) print(f'> atbash(): {timeit("atbash(printable)" , setup=snake_case )} seconds' ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(F"{example} encrypted in atbash: {atbash(example)}") benchmark()
85
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed __lowerCAmelCase : List[str] ='true' def _UpperCamelCase ( lowercase__ , lowercase__=82 , lowercase__=16 ): set_seed(42 ) __SCREAMING_SNAKE_CASE : Optional[int] = RegressionModel() __SCREAMING_SNAKE_CASE : Optional[int] = deepcopy(lowercase__ ) __SCREAMING_SNAKE_CASE : Any = RegressionDataset(length=lowercase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = DataLoader(lowercase__ , batch_size=lowercase__ ) model.to(accelerator.device ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[Any] = accelerator.prepare(lowercase__ , lowercase__ ) return model, ddp_model, dataloader def _UpperCamelCase ( lowercase__ , lowercase__=False ): __SCREAMING_SNAKE_CASE : Optional[Any] = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) __SCREAMING_SNAKE_CASE : str = load_dataset('''glue''' , '''mrpc''' , split='''validation''' ) def tokenize_function(lowercase__ ): __SCREAMING_SNAKE_CASE : Dict = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=lowercase__ , max_length=lowercase__ ) return outputs with accelerator.main_process_first(): __SCREAMING_SNAKE_CASE : Tuple = dataset.map( lowercase__ , batched=lowercase__ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) __SCREAMING_SNAKE_CASE : List[Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(lowercase__ ): if use_longest: return tokenizer.pad(lowercase__ , padding='''longest''' , return_tensors='''pt''' ) return tokenizer.pad(lowercase__ , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return DataLoader(lowercase__ , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=16 ) def _UpperCamelCase ( lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : str = Accelerator(dispatch_batches=lowercase__ , split_batches=lowercase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = get_dataloader(lowercase__ , not dispatch_batches ) __SCREAMING_SNAKE_CASE : List[str] = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''' , return_dict=lowercase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = accelerator.prepare(lowercase__ , lowercase__ ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def _UpperCamelCase ( lowercase__ , lowercase__ , lowercase__ ): __SCREAMING_SNAKE_CASE : List[str] = [] for batch in dataloader: __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[Any] = batch.values() with torch.no_grad(): __SCREAMING_SNAKE_CASE : Dict = model(lowercase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : str = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : str = [], [] for logit, targ in logits_and_targets: logits.append(lowercase__ ) targs.append(lowercase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Any = torch.cat(lowercase__ ), torch.cat(lowercase__ ) return logits, targs def _UpperCamelCase ( lowercase__ , lowercase__=82 , lowercase__=False , lowercase__=False , lowercase__=16 ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Tuple = get_basic_setup(lowercase__ , lowercase__ , lowercase__ ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = generate_predictions(lowercase__ , lowercase__ , lowercase__ ) assert ( len(lowercase__ ) == num_samples ), F'''Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(lowercase__ )}''' def _UpperCamelCase ( lowercase__ = False , lowercase__ = False ): __SCREAMING_SNAKE_CASE : Optional[Any] = evaluate.load('''glue''' , '''mrpc''' ) __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[Any] = get_mrpc_setup(lowercase__ , lowercase__ ) # First do baseline __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Optional[Any] = setup['''no'''] model.to(lowercase__ ) model.eval() for batch in dataloader: batch.to(lowercase__ ) with torch.inference_mode(): __SCREAMING_SNAKE_CASE : Dict = model(**lowercase__ ) __SCREAMING_SNAKE_CASE : Dict = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=lowercase__ , references=batch['''labels'''] ) __SCREAMING_SNAKE_CASE : int = metric.compute() # Then do distributed __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): __SCREAMING_SNAKE_CASE : int = model(**lowercase__ ) __SCREAMING_SNAKE_CASE : str = outputs.logits.argmax(dim=-1 ) __SCREAMING_SNAKE_CASE : Any = batch['''labels'''] __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Dict = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=lowercase__ , references=lowercase__ ) __SCREAMING_SNAKE_CASE : List[Any] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), F'''Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n''' def _UpperCamelCase ( ): __SCREAMING_SNAKE_CASE : Dict = Accelerator(split_batches=lowercase__ , dispatch_batches=lowercase__ ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`''' ) test_mrpc(lowercase__ , lowercase__ ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __SCREAMING_SNAKE_CASE : List[Any] = Accelerator(split_batches=lowercase__ , dispatch_batches=lowercase__ ) if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99''' ) test_torch_metrics(lowercase__ , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) __SCREAMING_SNAKE_CASE : Tuple = Accelerator() test_torch_metrics(lowercase__ , 512 ) accelerator.state._reset_state() def _UpperCamelCase ( lowercase__ ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
9
0
"""simple docstring""" from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCamelCase__ ( a__ ): """simple docstring""" _SCREAMING_SNAKE_CASE = ["vqvae"] def __init__( self : Optional[int] , SCREAMING_SNAKE_CASE_ : Tuple , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : List[str] , ): super().__init__() self.register_modules(unet=SCREAMING_SNAKE_CASE_ , scheduler=SCREAMING_SNAKE_CASE_ , mel=SCREAMING_SNAKE_CASE_ , vqvae=SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE__ ( self : Dict ): return 5_0 if isinstance(self.scheduler , SCREAMING_SNAKE_CASE_ ) else 1_0_0_0 @torch.no_grad() def __call__( self : List[str] , SCREAMING_SNAKE_CASE_ : Optional[int] = 1 , SCREAMING_SNAKE_CASE_ : str = None , SCREAMING_SNAKE_CASE_ : List[Any] = None , SCREAMING_SNAKE_CASE_ : List[Any] = 0 , SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 , SCREAMING_SNAKE_CASE_ : Optional[int] = None , SCREAMING_SNAKE_CASE_ : Optional[Any] = None , SCREAMING_SNAKE_CASE_ : Dict = 0 , SCREAMING_SNAKE_CASE_ : int = 0 , SCREAMING_SNAKE_CASE_ : Optional[Any] = None , SCREAMING_SNAKE_CASE_ : Tuple = 0 , SCREAMING_SNAKE_CASE_ : Tuple = None , SCREAMING_SNAKE_CASE_ : Tuple = None , SCREAMING_SNAKE_CASE_ : Tuple=True , ): lowerCAmelCase_ : Any = steps or self.get_default_steps() self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : Optional[int] = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: lowerCAmelCase_ : List[Any] = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: lowerCAmelCase_ : Optional[int] = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=SCREAMING_SNAKE_CASE_ , device=self.device , ) lowerCAmelCase_ : Dict = noise lowerCAmelCase_ : Optional[Any] = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : Optional[Any] = self.mel.audio_slice_to_image(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : Any = np.frombuffer(input_image.tobytes() , dtype='uint8' ).reshape( (input_image.height, input_image.width) ) lowerCAmelCase_ : Any = (input_image / 2_5_5) * 2 - 1 lowerCAmelCase_ : List[str] = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: lowerCAmelCase_ : List[str] = self.vqvae.encode(torch.unsqueeze(SCREAMING_SNAKE_CASE_ , 0 ) ).latent_dist.sample( generator=SCREAMING_SNAKE_CASE_ )[0] lowerCAmelCase_ : int = self.vqvae.config.scaling_factor * input_images if start_step > 0: lowerCAmelCase_ : int = self.scheduler.add_noise(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , self.scheduler.timesteps[start_step - 1] ) lowerCAmelCase_ : Tuple = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) lowerCAmelCase_ : Union[str, Any] = int(mask_start_secs * pixels_per_second ) lowerCAmelCase_ : Any = int(mask_end_secs * pixels_per_second ) lowerCAmelCase_ : Tuple = self.scheduler.add_noise(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase_ : List[str] = self.unet(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )['sample'] else: lowerCAmelCase_ : int = self.unet(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )['sample'] if isinstance(self.scheduler , SCREAMING_SNAKE_CASE_ ): lowerCAmelCase_ : Dict = self.scheduler.step( model_output=SCREAMING_SNAKE_CASE_ , timestep=SCREAMING_SNAKE_CASE_ , sample=SCREAMING_SNAKE_CASE_ , eta=SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , )['prev_sample'] else: lowerCAmelCase_ : str = self.scheduler.step( model_output=SCREAMING_SNAKE_CASE_ , timestep=SCREAMING_SNAKE_CASE_ , sample=SCREAMING_SNAKE_CASE_ , generator=SCREAMING_SNAKE_CASE_ , )['prev_sample'] if mask is not None: if mask_start > 0: lowerCAmelCase_ : str = mask[:, step, :, :mask_start] if mask_end > 0: lowerCAmelCase_ : Dict = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance lowerCAmelCase_ : str = 1 / self.vqvae.config.scaling_factor * images lowerCAmelCase_ : Union[str, Any] = self.vqvae.decode(SCREAMING_SNAKE_CASE_ )['sample'] lowerCAmelCase_ : int = (images / 2 + 0.5).clamp(0 , 1 ) lowerCAmelCase_ : Any = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() lowerCAmelCase_ : Dict = (images * 2_5_5).round().astype('uint8' ) lowerCAmelCase_ : List[str] = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(SCREAMING_SNAKE_CASE_ , mode='RGB' ).convert('L' ) for _ in images) ) lowerCAmelCase_ : Any = [self.mel.image_to_audio(SCREAMING_SNAKE_CASE_ ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(SCREAMING_SNAKE_CASE_ )[:, np.newaxis, :] ) , **ImagePipelineOutput(SCREAMING_SNAKE_CASE_ ) ) @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( self : List[Any] , SCREAMING_SNAKE_CASE_ : Dict , SCREAMING_SNAKE_CASE_ : Union[str, Any] = 5_0 ): assert isinstance(self.scheduler , SCREAMING_SNAKE_CASE_ ) self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase_ : List[Any] = np.array( [np.frombuffer(image.tobytes() , dtype='uint8' ).reshape((1, image.height, image.width) ) for image in images] ) lowerCAmelCase_ : str = (sample / 2_5_5) * 2 - 1 lowerCAmelCase_ : Optional[Any] = torch.Tensor(SCREAMING_SNAKE_CASE_ ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): lowerCAmelCase_ : str = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps lowerCAmelCase_ : Any = self.scheduler.alphas_cumprod[t] lowerCAmelCase_ : str = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) lowerCAmelCase_ : str = 1 - alpha_prod_t lowerCAmelCase_ : int = self.unet(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )['sample'] lowerCAmelCase_ : str = (1 - alpha_prod_t_prev) ** 0.5 * model_output lowerCAmelCase_ : List[str] = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) lowerCAmelCase_ : int = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE_ : str , SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : Optional[int] ): lowerCAmelCase_ : List[str] = acos(torch.dot(torch.flatten(SCREAMING_SNAKE_CASE_ ) , torch.flatten(SCREAMING_SNAKE_CASE_ ) ) / torch.norm(SCREAMING_SNAKE_CASE_ ) / torch.norm(SCREAMING_SNAKE_CASE_ ) ) return sin((1 - alpha) * theta ) * xa / sin(SCREAMING_SNAKE_CASE_ ) + sin(alpha * theta ) * xa / sin(SCREAMING_SNAKE_CASE_ )
365
"""simple docstring""" import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home lowercase__ : int = HUGGINGFACE_HUB_CACHE lowercase__ : Tuple = """config.json""" lowercase__ : Union[str, Any] = """diffusion_pytorch_model.bin""" lowercase__ : List[Any] = """diffusion_flax_model.msgpack""" lowercase__ : List[str] = """model.onnx""" lowercase__ : List[Any] = """diffusion_pytorch_model.safetensors""" lowercase__ : Dict = """weights.pb""" lowercase__ : List[Any] = """https://huggingface.co""" lowercase__ : List[Any] = default_cache_path lowercase__ : Tuple = """diffusers_modules""" lowercase__ : Tuple = os.getenv("""HF_MODULES_CACHE""", os.path.join(hf_cache_home, """modules""")) lowercase__ : List[str] = ["""fp16""", """non-ema"""] lowercase__ : Optional[Any] = """.self_attn"""
289
0
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCamelCase : '''simple docstring''' def __init__( self , __a , __a=2 , __a=8 , __a=True , __a=True , __a=True , __a=True , __a=99 , __a=16 , __a=5 , __a=2 , __a=36 , __a="gelu" , __a=0.0 , __a=0.0 , __a=5_12 , __a=16 , __a=2 , __a=0.0_2 , __a=3 , __a=4 , __a=None , ): __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_input_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_labels __lowerCAmelCase = num_choices __lowerCAmelCase = scope def snake_case ( self ): __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_input_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = None __lowerCAmelCase = None __lowerCAmelCase = None if self.use_labels: __lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) __lowerCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case ( self ): return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__a , initializer_range=self.initializer_range , ) def snake_case ( self ): __lowerCAmelCase = self.get_config() __lowerCAmelCase = 3_00 return config def snake_case ( self ): ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = self.prepare_config_and_inputs() __lowerCAmelCase = True __lowerCAmelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = MraModel(config=__a ) model.to(__a ) model.eval() __lowerCAmelCase = model(__a , attention_mask=__a , token_type_ids=__a ) __lowerCAmelCase = model(__a , token_type_ids=__a ) __lowerCAmelCase = model(__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a , __a , __a , ): __lowerCAmelCase = True __lowerCAmelCase = MraModel(__a ) model.to(__a ) model.eval() __lowerCAmelCase = model( __a , attention_mask=__a , token_type_ids=__a , encoder_hidden_states=__a , encoder_attention_mask=__a , ) __lowerCAmelCase = model( __a , attention_mask=__a , token_type_ids=__a , encoder_hidden_states=__a , ) __lowerCAmelCase = model(__a , attention_mask=__a , token_type_ids=__a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = MraForMaskedLM(config=__a ) model.to(__a ) model.eval() __lowerCAmelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = MraForQuestionAnswering(config=__a ) model.to(__a ) model.eval() __lowerCAmelCase = model( __a , attention_mask=__a , token_type_ids=__a , start_positions=__a , end_positions=__a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_labels __lowerCAmelCase = MraForSequenceClassification(__a ) model.to(__a ) model.eval() __lowerCAmelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_labels __lowerCAmelCase = MraForTokenClassification(config=__a ) model.to(__a ) model.eval() __lowerCAmelCase = model(__a , attention_mask=__a , token_type_ids=__a , labels=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case ( self , __a , __a , __a , __a , __a , __a , __a ): __lowerCAmelCase = self.num_choices __lowerCAmelCase = MraForMultipleChoice(config=__a ) model.to(__a ) model.eval() __lowerCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowerCAmelCase = model( __a , attention_mask=__a , token_type_ids=__a , labels=__a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case ( self ): __lowerCAmelCase = self.prepare_config_and_inputs() ( ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ( __lowerCAmelCase ) , ) = config_and_inputs __lowerCAmelCase = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _UpperCamelCase ( lowerCAmelCase__ ,unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : List[Any] =( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __UpperCAmelCase : str =False __UpperCAmelCase : int =False __UpperCAmelCase : Optional[Any] =False __UpperCAmelCase : Tuple =False __UpperCAmelCase : List[Any] =() def snake_case ( self ): __lowerCAmelCase = MraModelTester(self ) __lowerCAmelCase = ConfigTester(self , config_class=__a , hidden_size=37 ) def snake_case ( self ): self.config_tester.run_common_tests() def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowerCAmelCase = type self.model_tester.create_and_check_model(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__a ) def snake_case ( self ): __lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__a ) @slow def snake_case ( self ): for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowerCAmelCase = MraModel.from_pretrained(__a ) self.assertIsNotNone(__a ) @unittest.skip(reason="MRA does not output attentions" ) def snake_case ( self ): return @require_torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def snake_case ( self ): __lowerCAmelCase = MraModel.from_pretrained("uw-madison/mra-base-512-4" ) __lowerCAmelCase = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): __lowerCAmelCase = model(__a )[0] __lowerCAmelCase = torch.Size((1, 2_56, 7_68) ) self.assertEqual(output.shape , __a ) __lowerCAmelCase = torch.tensor( [[[-0.0_1_4_0, 0.0_8_3_0, -0.0_3_8_1], [0.1_5_4_6, 0.1_4_0_2, 0.0_2_2_0], [0.1_1_6_2, 0.0_8_5_1, 0.0_1_6_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __a , atol=1e-4 ) ) @slow def snake_case ( self ): __lowerCAmelCase = MraForMaskedLM.from_pretrained("uw-madison/mra-base-512-4" ) __lowerCAmelCase = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): __lowerCAmelCase = model(__a )[0] __lowerCAmelCase = 5_02_65 __lowerCAmelCase = torch.Size((1, 2_56, vocab_size) ) self.assertEqual(output.shape , __a ) __lowerCAmelCase = torch.tensor( [[[9.2_5_9_5, -3.6_0_3_8, 1_1.8_8_1_9], [9.3_8_6_9, -3.2_6_9_3, 1_1.0_9_5_6], [1_1.8_5_2_4, -3.4_9_3_8, 1_3.1_2_1_0]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __a , atol=1e-4 ) ) @slow def snake_case ( self ): __lowerCAmelCase = MraForMaskedLM.from_pretrained("uw-madison/mra-base-4096-8-d3" ) __lowerCAmelCase = torch.arange(40_96 ).unsqueeze(0 ) with torch.no_grad(): __lowerCAmelCase = model(__a )[0] __lowerCAmelCase = 5_02_65 __lowerCAmelCase = torch.Size((1, 40_96, vocab_size) ) self.assertEqual(output.shape , __a ) __lowerCAmelCase = torch.tensor( [[[5.4_7_8_9, -2.3_5_6_4, 7.5_0_6_4], [7.9_0_6_7, -1.3_3_6_9, 9.9_6_6_8], [9.0_7_1_2, -1.8_1_0_6, 7.0_3_8_0]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __a , atol=1e-4 ) )
57
import warnings from functools import wraps from typing import Callable def UpperCAmelCase ( lowercase ): """simple docstring""" @wraps(lowercase ) def _inner_fn(*lowercase , **lowercase ): warnings.warn( (F"'{fn.__name__}' is experimental and might be subject to breaking changes in the future.") , lowercase , ) return fn(*lowercase , **lowercase ) return _inner_fn
210
0
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __snake_case = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class __lowerCamelCase ( a__ , unittest.TestCase ): '''simple docstring''' A_ : List[str] = XGLMTokenizer A_ : Optional[Any] = XGLMTokenizerFast A_ : Any = True A_ : str = True def _UpperCAmelCase ( self ) -> Tuple: super().setUp() # We have a SentencePiece fixture for testing _a = XGLMTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def _UpperCAmelCase ( self ) -> str: _a = '''<pad>''' _a = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def _UpperCAmelCase ( self ) -> Optional[int]: _a = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(len(__UpperCAmelCase ) , 1008 ) def _UpperCAmelCase ( self ) -> List[str]: self.assertEqual(self.get_tokenizer().vocab_size , 1008 ) def _UpperCAmelCase ( self ) -> str: _a = XGLMTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) _a = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__UpperCAmelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) _a = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _a = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) _a = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def _UpperCAmelCase ( self ) -> List[Any]: return XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) def _UpperCAmelCase ( self ) -> Dict: with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__UpperCAmelCase , f.name ) _a = XGLMTokenizer(f.name , keep_accents=__UpperCAmelCase ) _a = pickle.dumps(__UpperCAmelCase ) pickle.loads(__UpperCAmelCase ) def _UpperCAmelCase ( self ) -> List[str]: if not self.test_rust_tokenizer: return _a = self.get_tokenizer() _a = self.get_rust_tokenizer() _a = '''I was born in 92000, and this is falsé.''' _a = tokenizer.tokenize(__UpperCAmelCase ) _a = rust_tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) _a = tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) _a = rust_tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) _a = self.get_rust_tokenizer() _a = tokenizer.encode(__UpperCAmelCase ) _a = rust_tokenizer.encode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) @slow def _UpperCAmelCase ( self ) -> Union[str, Any]: _a = '''Hello World!''' _a = [2, 31227, 4447, 35] self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def _UpperCAmelCase ( self ) -> Tuple: _a = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth''' ) # fmt: off _a = [2, 1018, 67, 11, 1988, 2617, 5631, 278, 11, 3407, 48, 71630, 28085, 4, 3234, 157, 13, 6, 5, 6, 4, 3526, 768, 15, 659, 57, 298, 3983, 864, 129, 21, 6, 5, 13675, 377, 652, 7580, 10341, 155, 2817, 422, 1666, 7, 1674, 53, 113, 202277, 17892, 33, 60, 87, 4, 3234, 157, 61, 2667, 52376, 19, 88, 23, 735] # fmt: on self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def _UpperCAmelCase ( self ) -> Optional[Any]: # fmt: off _a = { '''input_ids''': [[2, 108825, 1163, 15, 88010, 473, 15898, 157, 13672, 1857, 312, 8, 238021, 1163, 53, 13672, 1857, 312, 8, 53283, 182396, 8, 18566, 16, 36733, 4101, 8, 230, 244017, 122553, 7, 15, 132597, 4, 293, 12511, 7610, 4, 3414, 132597, 9, 4, 32361, 362, 4, 734, 28512, 32569, 18, 4, 32361, 26096, 14982, 73, 18715, 21433, 235261, 15, 492, 12427, 16, 53, 18715, 21433, 65454, 15, 23659, 563, 16, 278, 597, 2843, 595, 7931, 182396, 64186, 22, 886, 595, 132981, 53, 25540, 3449, 43982, 39901, 5951, 878, 330, 4, 27694, 80269, 312, 53, 6517, 11780, 611, 20408, 5], [2, 6, 132597, 67, 42897, 33, 592, 8, 163729, 25540, 361, 136997, 109514, 173230, 7, 501, 60, 102913, 196, 5631, 235, 63243, 473, 6, 231757, 74, 5277, 7905, 53, 3095, 37317, 22, 454, 183874, 5], [2, 268, 31298, 46530, 6, 132935, 43831, 7, 597, 32, 24, 3688, 9865, 5]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name='''facebook/xglm-564M''' , padding=__UpperCAmelCase , )
153
"""simple docstring""" import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class __lowerCamelCase : '''simple docstring''' @staticmethod def _UpperCAmelCase ( *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: pass def A_ ( _lowerCAmelCase : Image ): """simple docstring""" _a = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' A_ : List[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> str: _a = DepthEstimationPipeline(model=__UpperCAmelCase , image_processor=__UpperCAmelCase ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Dict: _a = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} , __UpperCAmelCase ) import datasets _a = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) _a = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] , __UpperCAmelCase , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def _UpperCAmelCase ( self ) -> Tuple: pass @slow @require_torch def _UpperCAmelCase ( self ) -> List[str]: _a = '''Intel/dpt-large''' _a = pipeline('''depth-estimation''' , model=__UpperCAmelCase ) _a = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) _a = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) , 29.304 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) , 2.662 ) @require_torch def _UpperCAmelCase ( self ) -> List[Any]: # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
153
1
from dataclasses import dataclass from typing import Optional import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .attention import BasicTransformerBlock from .modeling_utils import ModelMixin @dataclass class _a (__magic_name__ ): '''simple docstring''' UpperCAmelCase__: torch.FloatTensor class _a (__magic_name__ , __magic_name__ ): '''simple docstring''' @register_to_config def __init__( self , A__ = 16 , A__ = 88 , A__ = None , A__ = None , A__ = 1 , A__ = 0.0 , A__ = 32 , A__ = None , A__ = False , A__ = None , A__ = "geglu" , A__ = True , A__ = True , ): super().__init__() A__ : Union[str, Any] = num_attention_heads A__ : Union[str, Any] = attention_head_dim A__ : Dict = num_attention_heads * attention_head_dim A__ : List[str] = in_channels A__ : List[str] = torch.nn.GroupNorm(num_groups=A__ , num_channels=A__ , eps=1e-6 , affine=A__ ) A__ : Any = nn.Linear(A__ , A__ ) # 3. Define transformers blocks A__ : str = nn.ModuleList( [ BasicTransformerBlock( A__ , A__ , A__ , dropout=A__ , cross_attention_dim=A__ , activation_fn=A__ , attention_bias=A__ , double_self_attention=A__ , norm_elementwise_affine=A__ , ) for d in range(A__ ) ] ) A__ : Dict = nn.Linear(A__ , A__ ) def __A ( self , A__ , A__=None , A__=None , A__=None , A__=1 , A__=None , A__ = True , ): A__ , A__ , A__ , A__ : Union[str, Any] = hidden_states.shape A__ : int = batch_frames // num_frames A__ : Dict = hidden_states A__ : int = hidden_states[None, :].reshape(A__ , A__ , A__ , A__ , A__ ) A__ : List[str] = hidden_states.permute(0 , 2 , 1 , 3 , 4 ) A__ : Optional[Any] = self.norm(A__ ) A__ : str = hidden_states.permute(0 , 3 , 4 , 2 , 1 ).reshape(batch_size * height * width , A__ , A__ ) A__ : List[Any] = self.proj_in(A__ ) # 2. Blocks for block in self.transformer_blocks: A__ : Dict = block( A__ , encoder_hidden_states=A__ , timestep=A__ , cross_attention_kwargs=A__ , class_labels=A__ , ) # 3. Output A__ : List[Any] = self.proj_out(A__ ) A__ : Dict = ( hidden_states[None, None, :] .reshape(A__ , A__ , A__ , A__ , A__ ) .permute(0 , 3 , 4 , 1 , 2 ) .contiguous() ) A__ : Tuple = hidden_states.reshape(A__ , A__ , A__ , A__ ) A__ : Union[str, Any] = hidden_states + residual if not return_dict: return (output,) return TransformerTemporalModelOutput(sample=A__ )
192
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch A_ : List[Any] = logging.get_logger(__name__) class _a : '''simple docstring''' def __init__( self , A__ = None , A__ = None , A__=None , A__=None ): if not conversation_id: A__ : List[Any] = uuid.uuida() if past_user_inputs is None: A__ : Dict = [] if generated_responses is None: A__ : int = [] A__ : uuid.UUID = conversation_id A__ : List[str] = past_user_inputs A__ : List[str] = generated_responses A__ : Optional[str] = text def __eq__( self , A__ ): if not isinstance(A__ , A__ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def __A ( self , A__ , A__ = False ): if self.new_user_input: if overwrite: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ F"""with: \"{text}\".""" ) A__ : str = text else: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ F"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: A__ : Tuple = text def __A ( self ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) A__ : Tuple = None def __A ( self , A__ ): self.generated_responses.append(A__ ) def __A ( self ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self ): A__ : Optional[Any] = F"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): A__ : str = """user""" if is_user else """bot""" output += F"""{name} >> {text} \n""" return output @add_end_docstrings( __magic_name__ , r''' min_length_for_response (`int`, *optional*, defaults to 32): The minimum length (in number of tokens) for a response. minimum_tokens (`int`, *optional*, defaults to 10): The minimum length of tokens to leave for a response. ''' , ) class _a (__magic_name__ ): '''simple docstring''' def __init__( self , *A__ , **A__ ): super().__init__(*A__ , **A__ ) if self.tokenizer.pad_token_id is None: A__ : Tuple = self.tokenizer.eos_token def __A ( self , A__=None , A__=None , A__=None , **A__ ): A__ : Tuple = {} A__ : List[str] = {} A__ : Union[str, Any] = {} if min_length_for_response is not None: A__ : str = min_length_for_response if minimum_tokens is not None: A__ : List[str] = minimum_tokens if "max_length" in generate_kwargs: A__ : List[Any] = generate_kwargs["""max_length"""] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: A__ : Optional[int] = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(A__ ) return preprocess_params, forward_params, postprocess_params def __call__( self , A__ , A__=0 , **A__ ): A__ : Optional[Any] = super().__call__(A__ , num_workers=A__ , **A__ ) if isinstance(A__ , A__ ) and len(A__ ) == 1: return outputs[0] return outputs def __A ( self , A__ , A__=32 ): if not isinstance(A__ , A__ ): raise ValueError("""ConversationalPipeline, expects Conversation as inputs""" ) if conversation.new_user_input is None: raise ValueError( F"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ """Add user inputs with the conversation's `add_user_input` method""" ) if hasattr(self.tokenizer , """_build_conversation_input_ids""" ): A__ : List[str] = self.tokenizer._build_conversation_input_ids(A__ ) else: # If the tokenizer cannot handle conversations, we default to only the old version A__ : Tuple = self._legacy_parse_and_tokenize(A__ ) if self.framework == "pt": A__ : List[str] = torch.LongTensor([input_ids] ) elif self.framework == "tf": A__ : Optional[Any] = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def __A ( self , A__ , A__=10 , **A__ ): A__ : List[Any] = generate_kwargs.get("""max_length""" , self.model.config.max_length ) A__ : Optional[int] = model_inputs["""input_ids"""].shape[1] if max_length - minimum_tokens < n: logger.warning(F"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) A__ : Dict = max_length - minimum_tokens A__ : Optional[int] = model_inputs["""input_ids"""][:, -trim:] if "attention_mask" in model_inputs: A__ : str = model_inputs["""attention_mask"""][:, -trim:] A__ : List[str] = model_inputs.pop("""conversation""" ) A__ : Dict = max_length A__ : str = self.model.generate(**A__ , **A__ ) if self.model.config.is_encoder_decoder: A__ : Union[str, Any] = 1 else: A__ : Optional[Any] = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def __A ( self , A__ , A__=True ): A__ : Dict = model_outputs["""output_ids"""] A__ : str = self.tokenizer.decode( output_ids[0] , skip_special_tokens=A__ , clean_up_tokenization_spaces=A__ , ) A__ : Optional[int] = model_outputs["""conversation"""] conversation.mark_processed() conversation.append_response(A__ ) return conversation def __A ( self , A__ ): A__ : str = self.tokenizer.eos_token_id A__ : Tuple = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(A__ , add_special_tokens=A__ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(A__ , add_special_tokens=A__ ) ) if len(A__ ) > self.tokenizer.model_max_length: A__ : str = input_ids[-self.tokenizer.model_max_length :] return input_ids
192
1
import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def __lowercase ( _UpperCamelCase, _UpperCamelCase=0.9_9_9, _UpperCamelCase="cosine", ) ->Tuple: """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(_UpperCamelCase ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_UpperCamelCase ): return math.exp(t * -1_2.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) lowercase : List[str] = [] for i in range(_UpperCamelCase ): lowercase : List[str] = i / num_diffusion_timesteps lowercase : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(_UpperCamelCase ) / alpha_bar_fn(_UpperCamelCase ), _UpperCamelCase ) ) return torch.tensor(_UpperCamelCase, dtype=torch.floataa ) class __SCREAMING_SNAKE_CASE ( A__ , A__ ): A : Any = [e.name for e in KarrasDiffusionSchedulers] A : Dict = 2 @register_to_config def __init__( self , SCREAMING_SNAKE_CASE__ = 1000 , SCREAMING_SNAKE_CASE__ = 0.00085 , SCREAMING_SNAKE_CASE__ = 0.012 , SCREAMING_SNAKE_CASE__ = "linear" , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = "epsilon" , SCREAMING_SNAKE_CASE__ = "linspace" , SCREAMING_SNAKE_CASE__ = 0 , ): if trained_betas is not None: lowercase : str = torch.tensor(SCREAMING_SNAKE_CASE__ , dtype=torch.floataa ) elif beta_schedule == "linear": lowercase : Union[str, Any] = torch.linspace(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. lowercase : Union[str, Any] = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , SCREAMING_SNAKE_CASE__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule lowercase : str = betas_for_alpha_bar(SCREAMING_SNAKE_CASE__ ) else: raise NotImplementedError(f"""{beta_schedule} does is not implemented for {self.__class__}""" ) lowercase : Optional[int] = 1.0 - self.betas lowercase : Union[str, Any] = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__=None ): if schedule_timesteps is None: lowercase : Union[str, Any] = self.timesteps lowercase : int = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: lowercase : List[Any] = 1 if len(SCREAMING_SNAKE_CASE__ ) > 1 else 0 else: lowercase : int = timestep.cpu().item() if torch.is_tensor(SCREAMING_SNAKE_CASE__ ) else timestep lowercase : Optional[Any] = self._index_counter[timestep_int] return indices[pos].item() @property def __lowerCamelCase ( self ): # standard deviation of the initial noise distribution if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ): lowercase : Optional[Any] = self.index_for_timestep(SCREAMING_SNAKE_CASE__ ) if self.state_in_first_order: lowercase : Any = self.sigmas[step_index] else: lowercase : Optional[int] = self.sigmas_interpol[step_index] lowercase : Union[str, Any] = sample / ((sigma**2 + 1) ** 0.5) return sample def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = None , ): lowercase : Any = num_inference_steps lowercase : Optional[Any] = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": lowercase : Dict = np.linspace(0 , num_train_timesteps - 1 , SCREAMING_SNAKE_CASE__ , dtype=SCREAMING_SNAKE_CASE__ )[::-1].copy() elif self.config.timestep_spacing == "leading": lowercase : int = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 lowercase : Union[str, Any] = (np.arange(0 , SCREAMING_SNAKE_CASE__ ) * step_ratio).round()[::-1].copy().astype(SCREAMING_SNAKE_CASE__ ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": lowercase : str = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 lowercase : Dict = (np.arange(SCREAMING_SNAKE_CASE__ , 0 , -step_ratio )).round().copy().astype(SCREAMING_SNAKE_CASE__ ) timesteps -= 1 else: raise ValueError( f"""{self.config.timestep_spacing} is not supported. Please make sure to choose one of 'linspace', 'leading' or 'trailing'.""" ) lowercase : int = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) lowercase : Optional[int] = torch.from_numpy(np.log(SCREAMING_SNAKE_CASE__ ) ).to(SCREAMING_SNAKE_CASE__ ) lowercase : List[str] = np.interp(SCREAMING_SNAKE_CASE__ , np.arange(0 , len(SCREAMING_SNAKE_CASE__ ) ) , SCREAMING_SNAKE_CASE__ ) lowercase : Optional[int] = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) lowercase : str = torch.from_numpy(SCREAMING_SNAKE_CASE__ ).to(device=SCREAMING_SNAKE_CASE__ ) # interpolate sigmas lowercase : int = sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() lowercase : Optional[Any] = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) lowercase : Optional[int] = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(SCREAMING_SNAKE_CASE__ ).startswith('''mps''' ): # mps does not support float64 lowercase : Any = torch.from_numpy(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ , dtype=torch.floataa ) else: lowercase : List[Any] = torch.from_numpy(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ ) # interpolate timesteps lowercase : Any = self.sigma_to_t(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ , dtype=timesteps.dtype ) lowercase : List[Any] = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() lowercase : Dict = torch.cat([timesteps[:1], interleaved_timesteps] ) lowercase : int = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter lowercase : Dict = defaultdict(SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ ): # get log sigma lowercase : Any = sigma.log() # get distribution lowercase : Optional[Any] = log_sigma - self.log_sigmas[:, None] # get sigmas range lowercase : List[Any] = dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) lowercase : str = low_idx + 1 lowercase : Union[str, Any] = self.log_sigmas[low_idx] lowercase : Union[str, Any] = self.log_sigmas[high_idx] # interpolate sigmas lowercase : Dict = (low - log_sigma) / (low - high) lowercase : Union[str, Any] = w.clamp(0 , 1 ) # transform interpolation to time range lowercase : List[str] = (1 - w) * low_idx + w * high_idx lowercase : Tuple = t.view(sigma.shape ) return t @property def __lowerCamelCase ( self ): return self.sample is None def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = True , ): lowercase : Optional[Any] = self.index_for_timestep(SCREAMING_SNAKE_CASE__ ) # advance index counter by 1 lowercase : Dict = timestep.cpu().item() if torch.is_tensor(SCREAMING_SNAKE_CASE__ ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: lowercase : Union[str, Any] = self.sigmas[step_index] lowercase : List[Any] = self.sigmas_interpol[step_index + 1] lowercase : Any = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method lowercase : Any = self.sigmas[step_index - 1] lowercase : List[Any] = self.sigmas_interpol[step_index] lowercase : Optional[int] = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API lowercase : Union[str, Any] = 0 lowercase : List[Any] = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": lowercase : List[Any] = sigma_hat if self.state_in_first_order else sigma_interpol lowercase : Any = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": lowercase : List[str] = sigma_hat if self.state_in_first_order else sigma_interpol lowercase : Any = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError('''prediction_type not implemented yet: sample''' ) else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`""" ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order lowercase : List[str] = (sample - pred_original_sample) / sigma_hat # 3. delta timestep lowercase : Union[str, Any] = sigma_interpol - sigma_hat # store for 2nd order step lowercase : Optional[Any] = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order lowercase : str = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep lowercase : str = sigma_next - sigma_hat lowercase : List[str] = self.sample lowercase : Optional[int] = None lowercase : str = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=SCREAMING_SNAKE_CASE__ ) def __lowerCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ): # Make sure sigmas and timesteps have the same device and dtype as original_samples lowercase : int = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(SCREAMING_SNAKE_CASE__ ): # mps does not support float64 lowercase : Tuple = self.timesteps.to(original_samples.device , dtype=torch.floataa ) lowercase : List[Any] = timesteps.to(original_samples.device , dtype=torch.floataa ) else: lowercase : List[str] = self.timesteps.to(original_samples.device ) lowercase : Any = timesteps.to(original_samples.device ) lowercase : Tuple = [self.index_for_timestep(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for t in timesteps] lowercase : Union[str, Any] = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): lowercase : Any = sigma.unsqueeze(-1 ) lowercase : Optional[Any] = original_samples + noise * sigma return noisy_samples def __len__( self ): return self.config.num_train_timesteps
173
from math import pi, sqrt, tan def __lowercase ( _UpperCamelCase ) ->float: """simple docstring""" if side_length < 0: raise ValueError('''surface_area_cube() only accepts non-negative values''' ) return 6 * side_length**2 def __lowercase ( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if length < 0 or breadth < 0 or height < 0: raise ValueError('''surface_area_cuboid() only accepts non-negative values''' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def __lowercase ( _UpperCamelCase ) ->float: """simple docstring""" if radius < 0: raise ValueError('''surface_area_sphere() only accepts non-negative values''' ) return 4 * pi * radius**2 def __lowercase ( _UpperCamelCase ) ->float: """simple docstring""" if radius < 0: raise ValueError('''surface_area_hemisphere() only accepts non-negative values''' ) return 3 * pi * radius**2 def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if radius < 0 or height < 0: raise ValueError('''surface_area_cone() only accepts non-negative values''' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def __lowercase ( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( '''surface_area_conical_frustum() only accepts non-negative values''' ) lowercase : str = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if radius < 0 or height < 0: raise ValueError('''surface_area_cylinder() only accepts non-negative values''' ) return 2 * pi * radius * (height + radius) def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if torus_radius < 0 or tube_radius < 0: raise ValueError('''surface_area_torus() only accepts non-negative values''' ) if torus_radius < tube_radius: raise ValueError( '''surface_area_torus() does not support spindle or self intersecting tori''' ) return 4 * pow(_UpperCamelCase, 2 ) * torus_radius * tube_radius def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if length < 0 or width < 0: raise ValueError('''area_rectangle() only accepts non-negative values''' ) return length * width def __lowercase ( _UpperCamelCase ) ->float: """simple docstring""" if side_length < 0: raise ValueError('''area_square() only accepts non-negative values''' ) return side_length**2 def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if base < 0 or height < 0: raise ValueError('''area_triangle() only accepts non-negative values''' ) return (base * height) / 2 def __lowercase ( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('''area_triangle_three_sides() only accepts non-negative values''' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('''Given three sides do not form a triangle''' ) lowercase : List[Any] = (sidea + sidea + sidea) / 2 lowercase : Union[str, Any] = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if base < 0 or height < 0: raise ValueError('''area_parallelogram() only accepts non-negative values''' ) return base * height def __lowercase ( _UpperCamelCase, _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if basea < 0 or basea < 0 or height < 0: raise ValueError('''area_trapezium() only accepts non-negative values''' ) return 1 / 2 * (basea + basea) * height def __lowercase ( _UpperCamelCase ) ->float: """simple docstring""" if radius < 0: raise ValueError('''area_circle() only accepts non-negative values''' ) return pi * radius**2 def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if radius_x < 0 or radius_y < 0: raise ValueError('''area_ellipse() only accepts non-negative values''' ) return pi * radius_x * radius_y def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if diagonal_a < 0 or diagonal_a < 0: raise ValueError('''area_rhombus() only accepts non-negative values''' ) return 1 / 2 * diagonal_a * diagonal_a def __lowercase ( _UpperCamelCase, _UpperCamelCase ) ->float: """simple docstring""" if not isinstance(_UpperCamelCase, _UpperCamelCase ) or sides < 3: raise ValueError( '''area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides''' ) elif length < 0: raise ValueError( '''area_reg_polygon() only accepts non-negative values as \ length of a side''' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print('''[DEMO] Areas of various geometric shapes: \n''') print(F'''Rectangle: {area_rectangle(10, 20) = }''') print(F'''Square: {area_square(10) = }''') print(F'''Triangle: {area_triangle(10, 10) = }''') print(F'''Triangle: {area_triangle_three_sides(5, 12, 13) = }''') print(F'''Parallelogram: {area_parallelogram(10, 20) = }''') print(F'''Rhombus: {area_rhombus(10, 20) = }''') print(F'''Trapezium: {area_trapezium(10, 20, 30) = }''') print(F'''Circle: {area_circle(20) = }''') print(F'''Ellipse: {area_ellipse(10, 20) = }''') print('''\nSurface Areas of various geometric shapes: \n''') print(F'''Cube: {surface_area_cube(20) = }''') print(F'''Cuboid: {surface_area_cuboid(10, 20, 30) = }''') print(F'''Sphere: {surface_area_sphere(20) = }''') print(F'''Hemisphere: {surface_area_hemisphere(20) = }''') print(F'''Cone: {surface_area_cone(10, 20) = }''') print(F'''Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }''') print(F'''Cylinder: {surface_area_cylinder(10, 20) = }''') print(F'''Torus: {surface_area_torus(20, 10) = }''') print(F'''Equilateral Triangle: {area_reg_polygon(3, 10) = }''') print(F'''Square: {area_reg_polygon(4, 10) = }''') print(F'''Reqular Pentagon: {area_reg_polygon(5, 10) = }''')
173
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __a : str = { """configuration_efficientnet""": [ """EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientNetConfig""", """EfficientNetOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a : Union[str, Any] = ["""EfficientNetImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __a : Union[str, Any] = [ """EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientNetForImageClassification""", """EfficientNetModel""", """EfficientNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys __a : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
210
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time __a : List[str] = Lock() def UpperCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ): """simple docstring""" global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(lowercase ) process_lock.release() # receive your right neighbor's value process_lock.acquire() __lowercase = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left __lowercase = min(lowercase , lowercase ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(lowercase ) process_lock.release() # receive your left neighbor's value process_lock.acquire() __lowercase = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right __lowercase = max(lowercase , lowercase ) # after all swaps are performed, send the values back to main result_pipe[1].send(lowercase ) def UpperCAmelCase ( lowercase ): """simple docstring""" __lowercase = [] __lowercase = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop __lowercase = Pipe() __lowercase = Pipe() process_array_.append( Process( target=lowercase , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) __lowercase = temp_rs __lowercase = temp_rr for i in range(1 , len(lowercase ) - 1 ): __lowercase = Pipe() __lowercase = Pipe() process_array_.append( Process( target=lowercase , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) __lowercase = temp_rs __lowercase = temp_rr process_array_.append( Process( target=lowercase , args=( len(lowercase ) - 1, arr[len(lowercase ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(lowercase ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(lowercase ) ): __lowercase = result_pipe[p][0].recv() process_array_[p].join() return arr def UpperCAmelCase ( ): """simple docstring""" __lowercase = list(range(10 , 0 , -1 ) ) print('''Initial List''' ) print(*lowercase ) __lowercase = odd_even_transposition(lowercase ) print('''Sorted List\n''' ) print(*lowercase ) if __name__ == "__main__": main()
210
1
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __snake_case ( lowerCamelCase__ ): @require_torch def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : List[Any] =''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' UpperCAmelCase : Tuple =''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' UpperCAmelCase : int =''' import socket def offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache UpperCAmelCase : Optional[int] ='''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(snake_case__ ) BertModel.from_pretrained(snake_case__ ) BertTokenizer.from_pretrained(snake_case__ ) pipeline(task='''fill-mask''' , model=snake_case__ ) # baseline - just load from_pretrained with normal network UpperCAmelCase : List[Any] =[sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed UpperCAmelCase : List[Any] =self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase : Optional[Any] ='''1''' UpperCAmelCase : List[Any] =subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Optional[Any] =''' from transformers import BertConfig, BertModel, BertTokenizer, pipeline ''' UpperCAmelCase : Any =''' mname = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) BertTokenizer.from_pretrained(mname) pipe = pipeline(task="fill-mask", model=mname) print("success") ''' UpperCAmelCase : Union[str, Any] =''' import socket def offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet") socket.socket = offline_socket ''' # Force fetching the files so that we can use the cache UpperCAmelCase : Union[str, Any] ='''hf-internal-testing/tiny-random-bert''' BertConfig.from_pretrained(snake_case__ ) BertModel.from_pretrained(snake_case__ ) BertTokenizer.from_pretrained(snake_case__ ) pipeline(task='''fill-mask''' , model=snake_case__ ) # baseline - just load from_pretrained with normal network UpperCAmelCase : Any =[sys.executable, '''-c''', '''\n'''.join([load, run, mock] )] # should succeed UpperCAmelCase : List[str] =self.get_env() UpperCAmelCase : Any =subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def UpperCAmelCase__ ( self ) -> List[Any]: '''simple docstring''' UpperCAmelCase : Union[str, Any] =''' from transformers import BertConfig, BertModel, BertTokenizer ''' UpperCAmelCase : int =''' mname = "hf-internal-testing/tiny-random-bert-sharded" BertConfig.from_pretrained(mname) BertModel.from_pretrained(mname) print("success") ''' UpperCAmelCase : int =''' import socket def offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled") socket.socket = offline_socket ''' # baseline - just load from_pretrained with normal network UpperCAmelCase : Dict =[sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed UpperCAmelCase : Any =self.get_env() UpperCAmelCase : List[Any] =subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # next emulate no network UpperCAmelCase : Optional[Any] =[sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase : int ='''1''' UpperCAmelCase : Optional[Any] =subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) @require_torch def UpperCAmelCase__ ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Dict =''' from transformers import pipeline ''' UpperCAmelCase : List[Any] =''' mname = "hf-internal-testing/tiny-random-bert" pipe = pipeline(model=mname) ''' UpperCAmelCase : Tuple =''' import socket def offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled") socket.socket = offline_socket ''' UpperCAmelCase : Optional[int] =self.get_env() UpperCAmelCase : int ='''1''' UpperCAmelCase : Optional[int] =[sys.executable, '''-c''', '''\n'''.join([load, mock, run] )] UpperCAmelCase : List[str] =subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( '''You cannot infer task automatically within `pipeline` when using offline mode''' , result.stderr.decode().replace('''\n''' , '''''' ) , ) @require_torch def UpperCAmelCase__ ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : Any =''' from transformers import AutoModel ''' UpperCAmelCase : Optional[Any] =''' mname = "hf-internal-testing/test_dynamic_model" AutoModel.from_pretrained(mname, trust_remote_code=True) print("success") ''' # baseline - just load from_pretrained with normal network UpperCAmelCase : Dict =[sys.executable, '''-c''', '''\n'''.join([load, run] )] # should succeed UpperCAmelCase : Optional[int] =self.get_env() UpperCAmelCase : Optional[Any] =subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files UpperCAmelCase : Any ='''1''' UpperCAmelCase : Dict =subprocess.run(snake_case__ , env=snake_case__ , check=snake_case__ , capture_output=snake_case__ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('''success''' , result.stdout.decode() )
78
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
78
1
import argparse import csv import logging import os import random import numpy as np import torch from torch.utils.data import DataLoader, RandomSampler, SequentialSampler, TensorDataset from tqdm import tqdm, trange from transformers import ( CONFIG_NAME, WEIGHTS_NAME, AdamW, OpenAIGPTDoubleHeadsModel, OpenAIGPTTokenizer, get_linear_schedule_with_warmup, ) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) _snake_case = logging.getLogger(__name__) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : List[str] = np.argmax(a__ , axis=1 ) return np.sum(outputs == labels ) def lowercase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' with open(a__ , encoding="utf_8" ) as f: lowerCamelCase : Optional[int] = csv.reader(a__ ) lowerCamelCase : List[str] = [] next(a__ ) # skip the first line for line in tqdm(a__ ): output.append((" ".join(line[1:5] ), line[5], line[6], int(line[-1] ) - 1) ) return output def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : List[Any] = [] for dataset in encoded_datasets: lowerCamelCase : Dict = len(a__ ) lowerCamelCase : str = np.zeros((n_batch, 2, input_len) , dtype=np.intaa ) lowerCamelCase : int = np.zeros((n_batch, 2) , dtype=np.intaa ) lowerCamelCase : Union[str, Any] = np.full((n_batch, 2, input_len) , fill_value=-100 , dtype=np.intaa ) lowerCamelCase : Any = np.zeros((n_batch,) , dtype=np.intaa ) for ( i, (story, conta, conta, mc_label), ) in enumerate(a__ ): lowerCamelCase : Dict = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] lowerCamelCase : List[str] = [start_token] + story[:cap_length] + [delimiter_token] + conta[:cap_length] + [clf_token] lowerCamelCase : Optional[Any] = with_conta lowerCamelCase : List[Any] = with_conta lowerCamelCase : Any = len(a__ ) - 1 lowerCamelCase : str = len(a__ ) - 1 lowerCamelCase : Dict = with_conta lowerCamelCase : Optional[Any] = with_conta lowerCamelCase : Any = mc_label lowerCamelCase : int = (input_ids, mc_token_ids, lm_labels, mc_labels) tensor_datasets.append(tuple(torch.tensor(a__ ) for t in all_inputs ) ) return tensor_datasets def lowercase_( ): '''simple docstring''' lowerCamelCase : int = argparse.ArgumentParser() parser.add_argument("--model_name" , type=a__ , default="openai-gpt" , help="pretrained model name" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_eval" , action="store_true" , help="Whether to run eval on the dev set." ) parser.add_argument( "--output_dir" , default=a__ , type=a__ , required=a__ , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument("--train_dataset" , type=a__ , default="" ) parser.add_argument("--eval_dataset" , type=a__ , default="" ) parser.add_argument("--seed" , type=a__ , default=42 ) parser.add_argument("--num_train_epochs" , type=a__ , default=3 ) parser.add_argument("--train_batch_size" , type=a__ , default=8 ) parser.add_argument("--eval_batch_size" , type=a__ , default=16 ) parser.add_argument("--adam_epsilon" , default=1E-8 , type=a__ , help="Epsilon for Adam optimizer." ) parser.add_argument("--max_grad_norm" , type=a__ , default=1 ) parser.add_argument( "--max_steps" , default=-1 , type=a__ , help=( "If > 0: set total number of training steps to perform. Override num_train_epochs." ) , ) parser.add_argument( "--gradient_accumulation_steps" , type=a__ , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--learning_rate" , type=a__ , default=6.25E-5 ) parser.add_argument("--warmup_steps" , default=0 , type=a__ , help="Linear warmup over warmup_steps." ) parser.add_argument("--lr_schedule" , type=a__ , default="warmup_linear" ) parser.add_argument("--weight_decay" , type=a__ , default=0.01 ) parser.add_argument("--lm_coef" , type=a__ , default=0.9 ) parser.add_argument("--n_valid" , type=a__ , default=374 ) parser.add_argument("--server_ip" , type=a__ , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=a__ , default="" , help="Can be used for distant debugging." ) lowerCamelCase : Optional[Any] = parser.parse_args() print(a__ ) if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=a__ ) ptvsd.wait_for_attach() random.seed(args.seed ) np.random.seed(args.seed ) torch.manual_seed(args.seed ) torch.cuda.manual_seed_all(args.seed ) lowerCamelCase : Any = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) lowerCamelCase : Union[str, Any] = torch.cuda.device_count() logger.info("device: {}, n_gpu {}".format(a__ , a__ ) ) if not args.do_train and not args.do_eval: raise ValueError("At least one of `do_train` or `do_eval` must be True." ) if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) # Load tokenizer and model # This loading functions also add new tokens and embeddings called `special tokens` # These new embeddings will be fine-tuned on the RocStories dataset lowerCamelCase : Any = ["_start_", "_delimiter_", "_classify_"] lowerCamelCase : str = OpenAIGPTTokenizer.from_pretrained(args.model_name ) tokenizer.add_tokens(a__ ) lowerCamelCase : List[Any] = tokenizer.convert_tokens_to_ids(a__ ) lowerCamelCase : str = OpenAIGPTDoubleHeadsModel.from_pretrained(args.model_name ) model.resize_token_embeddings(len(a__ ) ) model.to(a__ ) # Load and encode the datasets def tokenize_and_encode(SCREAMING_SNAKE_CASE_ ): if isinstance(a__ , a__ ): return tokenizer.convert_tokens_to_ids(tokenizer.tokenize(a__ ) ) elif isinstance(a__ , a__ ): return obj return [tokenize_and_encode(a__ ) for o in obj] logger.info("Encoding dataset..." ) lowerCamelCase : Any = load_rocstories_dataset(args.train_dataset ) lowerCamelCase : Union[str, Any] = load_rocstories_dataset(args.eval_dataset ) lowerCamelCase : Dict = (train_dataset, eval_dataset) lowerCamelCase : List[Any] = tokenize_and_encode(a__ ) # Compute the max input length for the Transformer lowerCamelCase : List[str] = model.config.n_positions // 2 - 2 lowerCamelCase : Union[str, Any] = max( len(story[:max_length] ) + max(len(conta[:max_length] ) , len(conta[:max_length] ) ) + 3 for dataset in encoded_datasets for story, conta, conta, _ in dataset ) lowerCamelCase : Any = min(a__ , model.config.n_positions ) # Max size of input for the pre-trained model # Prepare inputs tensors and dataloaders lowerCamelCase : Union[str, Any] = pre_process_datasets(a__ , a__ , a__ , *a__ ) lowerCamelCase , lowerCamelCase : Union[str, Any] = tensor_datasets[0], tensor_datasets[1] lowerCamelCase : Optional[int] = TensorDataset(*a__ ) lowerCamelCase : int = RandomSampler(a__ ) lowerCamelCase : Union[str, Any] = DataLoader(a__ , sampler=a__ , batch_size=args.train_batch_size ) lowerCamelCase : int = TensorDataset(*a__ ) lowerCamelCase : str = SequentialSampler(a__ ) lowerCamelCase : Optional[Any] = DataLoader(a__ , sampler=a__ , batch_size=args.eval_batch_size ) # Prepare optimizer if args.do_train: if args.max_steps > 0: lowerCamelCase : Any = args.max_steps lowerCamelCase : Optional[Any] = args.max_steps // (len(a__ ) // args.gradient_accumulation_steps) + 1 else: lowerCamelCase : List[str] = len(a__ ) // args.gradient_accumulation_steps * args.num_train_epochs lowerCamelCase : int = list(model.named_parameters() ) lowerCamelCase : Tuple = ["bias", "LayerNorm.bias", "LayerNorm.weight"] lowerCamelCase : Any = [ { "params": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay )], "weight_decay": args.weight_decay, }, {"params": [p for n, p in param_optimizer if any(nd in n for nd in no_decay )], "weight_decay": 0.0}, ] lowerCamelCase : int = AdamW(a__ , lr=args.learning_rate , eps=args.adam_epsilon ) lowerCamelCase : Optional[Any] = get_linear_schedule_with_warmup( a__ , num_warmup_steps=args.warmup_steps , num_training_steps=a__ ) if args.do_train: lowerCamelCase , lowerCamelCase , lowerCamelCase : Dict = 0, 0, None model.train() for _ in trange(int(args.num_train_epochs ) , desc="Epoch" ): lowerCamelCase : str = 0 lowerCamelCase : int = 0 lowerCamelCase : Optional[Any] = tqdm(a__ , desc="Training" ) for step, batch in enumerate(a__ ): lowerCamelCase : Optional[Any] = tuple(t.to(a__ ) for t in batch ) lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : str = batch lowerCamelCase : Optional[int] = model(a__ , mc_token_ids=a__ , lm_labels=a__ , mc_labels=a__ ) lowerCamelCase : Dict = args.lm_coef * losses[0] + losses[1] loss.backward() optimizer.step() scheduler.step() optimizer.zero_grad() tr_loss += loss.item() lowerCamelCase : Any = ( loss.item() if exp_average_loss is None else 0.7 * exp_average_loss + 0.3 * loss.item() ) nb_tr_steps += 1 lowerCamelCase : List[Any] = "Training loss: {:.2e} lr: {:.2e}".format(a__ , scheduler.get_lr()[0] ) # Save a trained model if args.do_train: # Save a trained model, configuration and tokenizer lowerCamelCase : Dict = model.module if hasattr(a__ , "module" ) else model # Only save the model itself # If we save using the predefined names, we can load using `from_pretrained` lowerCamelCase : Any = os.path.join(args.output_dir , a__ ) lowerCamelCase : List[str] = os.path.join(args.output_dir , a__ ) torch.save(model_to_save.state_dict() , a__ ) model_to_save.config.to_json_file(a__ ) tokenizer.save_vocabulary(args.output_dir ) # Load a trained model and vocabulary that you have fine-tuned lowerCamelCase : List[str] = OpenAIGPTDoubleHeadsModel.from_pretrained(args.output_dir ) lowerCamelCase : Dict = OpenAIGPTTokenizer.from_pretrained(args.output_dir ) model.to(a__ ) if args.do_eval: model.eval() lowerCamelCase , lowerCamelCase : Any = 0, 0 lowerCamelCase , lowerCamelCase : Optional[int] = 0, 0 for batch in tqdm(a__ , desc="Evaluating" ): lowerCamelCase : Optional[Any] = tuple(t.to(a__ ) for t in batch ) lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : List[str] = batch with torch.no_grad(): lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase : Any = model( a__ , mc_token_ids=a__ , lm_labels=a__ , mc_labels=a__ ) lowerCamelCase : str = mc_logits.detach().cpu().numpy() lowerCamelCase : Union[str, Any] = mc_labels.to("cpu" ).numpy() lowerCamelCase : Optional[Any] = accuracy(a__ , a__ ) eval_loss += mc_loss.mean().item() eval_accuracy += tmp_eval_accuracy nb_eval_examples += input_ids.size(0 ) nb_eval_steps += 1 lowerCamelCase : Any = eval_loss / nb_eval_steps lowerCamelCase : Dict = eval_accuracy / nb_eval_examples lowerCamelCase : int = tr_loss / nb_tr_steps if args.do_train else None lowerCamelCase : Optional[Any] = {"eval_loss": eval_loss, "eval_accuracy": eval_accuracy, "train_loss": train_loss} lowerCamelCase : int = os.path.join(args.output_dir , "eval_results.txt" ) with open(a__ , "w" ) as writer: logger.info("***** Eval results *****" ) for key in sorted(result.keys() ): logger.info(" %s = %s" , a__ , str(result[key] ) ) writer.write("%s = %s\n" % (key, str(result[key] )) ) if __name__ == "__main__": main()
283
import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowercase_ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): A__ : Any = DanceDiffusionPipeline A__ : Any = UNCONDITIONAL_AUDIO_GENERATION_PARAMS A__ : List[Any] = PipelineTesterMixin.required_optional_params - { """callback""", """latents""", """callback_steps""", """output_type""", """num_images_per_prompt""", } A__ : Dict = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS A__ : str = False A__ : Any = False def lowerCamelCase_ ( self ): """simple docstring""" torch.manual_seed(0 ) UpperCamelCase_ = UNetaDModel( block_out_channels=(3_2, 3_2, 6_4) , extra_in_channels=1_6 , sample_size=5_1_2 , sample_rate=1_6_0_0_0 , in_channels=2 , out_channels=2 , flip_sin_to_cos=__UpperCamelCase , use_timestep_embedding=__UpperCamelCase , time_embedding_type="""fourier""" , mid_block_type="""UNetMidBlock1D""" , down_block_types=("""DownBlock1DNoSkip""", """DownBlock1D""", """AttnDownBlock1D""") , up_block_types=("""AttnUpBlock1D""", """UpBlock1D""", """UpBlock1DNoSkip""") , ) UpperCamelCase_ = IPNDMScheduler() UpperCamelCase_ = { """unet""": unet, """scheduler""": scheduler, } return components def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase=0 ): """simple docstring""" if str(__UpperCamelCase ).startswith("""mps""" ): UpperCamelCase_ = torch.manual_seed(__UpperCamelCase ) else: UpperCamelCase_ = torch.Generator(device=__UpperCamelCase ).manual_seed(__UpperCamelCase ) UpperCamelCase_ = { """batch_size""": 1, """generator""": generator, """num_inference_steps""": 4, } return inputs def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = """cpu""" # ensure determinism for the device-dependent torch.Generator UpperCamelCase_ = self.get_dummy_components() UpperCamelCase_ = DanceDiffusionPipeline(**__UpperCamelCase ) UpperCamelCase_ = pipe.to(__UpperCamelCase ) pipe.set_progress_bar_config(disable=__UpperCamelCase ) UpperCamelCase_ = self.get_dummy_inputs(__UpperCamelCase ) UpperCamelCase_ = pipe(**__UpperCamelCase ) UpperCamelCase_ = output.audios UpperCamelCase_ = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) UpperCamelCase_ = np.array([-0.7_265, 1.0_000, -0.8_388, 0.1_175, 0.9_498, -1.0_000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def lowerCamelCase_ ( self ): """simple docstring""" return super().test_save_load_local() @skip_mps def lowerCamelCase_ ( self ): """simple docstring""" return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) @skip_mps def lowerCamelCase_ ( self ): """simple docstring""" return super().test_save_load_optional_components() @skip_mps def lowerCamelCase_ ( self ): """simple docstring""" return super().test_attention_slicing_forward_pass() def lowerCamelCase_ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class lowercase_ ( unittest.TestCase ): def lowerCamelCase_ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = torch_device UpperCamelCase_ = DanceDiffusionPipeline.from_pretrained("""harmonai/maestro-150k""" ) UpperCamelCase_ = pipe.to(__UpperCamelCase ) pipe.set_progress_bar_config(disable=__UpperCamelCase ) UpperCamelCase_ = torch.manual_seed(0 ) UpperCamelCase_ = pipe(generator=__UpperCamelCase , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) UpperCamelCase_ = output.audios UpperCamelCase_ = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCamelCase_ = np.array([-0.0_192, -0.0_231, -0.0_318, -0.0_059, 0.0_002, -0.0_020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2 def lowerCamelCase_ ( self ): """simple docstring""" UpperCamelCase_ = torch_device UpperCamelCase_ = DanceDiffusionPipeline.from_pretrained("""harmonai/maestro-150k""" , torch_dtype=torch.floataa ) UpperCamelCase_ = pipe.to(__UpperCamelCase ) pipe.set_progress_bar_config(disable=__UpperCamelCase ) UpperCamelCase_ = torch.manual_seed(0 ) UpperCamelCase_ = pipe(generator=__UpperCamelCase , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) UpperCamelCase_ = output.audios UpperCamelCase_ = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCamelCase_ = np.array([-0.0_367, -0.0_488, -0.0_771, -0.0_525, -0.0_444, -0.0_341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1e-2
122
0
'''simple docstring''' import qiskit def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> int: __lowerCamelCase = qiskit.Aer.get_backend('''aer_simulator''' ) # Create a Quantum Circuit acting on the q register __lowerCamelCase = qiskit.QuantumCircuit(_a , _a ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator __lowerCamelCase = qiskit.execute(_a , _a , shots=10_00 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(_a ) if __name__ == "__main__": print(f'Total count for various states are: {single_qubit_measure(1, 1)}')
365
'''simple docstring''' import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class a__ ( UpperCAmelCase__ , unittest.TestCase ): lowerCamelCase : Optional[Any] =LongformerTokenizer lowerCamelCase : Optional[Any] =True lowerCamelCase : List[str] =LongformerTokenizerFast lowerCamelCase : Union[str, Any] =True def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowerCamelCase = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', ] __lowerCamelCase = dict(zip(a , range(len(a ) ) ) ) __lowerCamelCase = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] __lowerCamelCase = {'''unk_token''': '''<unk>'''} __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) __lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(a ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(a ) ) def SCREAMING_SNAKE_CASE__ ( self : int , **a : int ): """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **a ) def SCREAMING_SNAKE_CASE__ ( self : str , **a : Dict ): """simple docstring""" kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname , **a ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , a : int ): """simple docstring""" __lowerCamelCase = '''lower newer''' __lowerCamelCase = '''lower newer''' return input_text, output_text def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): """simple docstring""" __lowerCamelCase = self.tokenizer_class(self.vocab_file , self.merges_file , **self.special_tokens_map ) __lowerCamelCase = '''lower newer''' __lowerCamelCase = ['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] __lowerCamelCase = tokenizer.tokenize(a ) # , add_prefix_space=True) self.assertListEqual(a , a ) __lowerCamelCase = tokens + [tokenizer.unk_token] __lowerCamelCase = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(a ) , a ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" __lowerCamelCase = self.get_tokenizer() self.assertListEqual(tokenizer.encode('''Hello world!''' , add_special_tokens=a ) , [0, 3_14_14, 2_32, 3_28, 2] ) self.assertListEqual( tokenizer.encode('''Hello world! cécé herlolip 418''' , add_special_tokens=a ) , [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2] , ) @slow def SCREAMING_SNAKE_CASE__ ( self : Any ): """simple docstring""" __lowerCamelCase = self.tokenizer_class.from_pretrained('''allenai/longformer-base-4096''' ) __lowerCamelCase = tokenizer.encode('''sequence builders''' , add_special_tokens=a ) __lowerCamelCase = tokenizer.encode('''multi-sequence build''' , add_special_tokens=a ) __lowerCamelCase = tokenizer.encode( '''sequence builders''' , add_special_tokens=a , add_prefix_space=a ) __lowerCamelCase = tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=a , add_prefix_space=a ) __lowerCamelCase = tokenizer.build_inputs_with_special_tokens(a ) __lowerCamelCase = tokenizer.build_inputs_with_special_tokens(a , a ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def SCREAMING_SNAKE_CASE__ ( self : List[str] ): """simple docstring""" __lowerCamelCase = self.get_tokenizer() __lowerCamelCase = '''Encode this sequence.''' __lowerCamelCase = tokenizer.byte_encoder[''' '''.encode('''utf-8''' )[0]] # Testing encoder arguments __lowerCamelCase = tokenizer.encode(a , add_special_tokens=a , add_prefix_space=a ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(a , a ) __lowerCamelCase = tokenizer.encode(a , add_special_tokens=a , add_prefix_space=a ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(a , a ) tokenizer.add_special_tokens({'''bos_token''': '''<s>'''} ) __lowerCamelCase = tokenizer.encode(a , add_special_tokens=a ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(a , a ) # Testing spaces after special tokens __lowerCamelCase = '''<mask>''' tokenizer.add_special_tokens( {'''mask_token''': AddedToken(a , lstrip=a , rstrip=a )} ) # mask token has a left space __lowerCamelCase = tokenizer.convert_tokens_to_ids(a ) __lowerCamelCase = '''Encode <mask> sequence''' __lowerCamelCase = '''Encode <mask>sequence''' __lowerCamelCase = tokenizer.encode(a ) __lowerCamelCase = encoded.index(a ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(a , a ) __lowerCamelCase = tokenizer.encode(a ) __lowerCamelCase = encoded.index(a ) __lowerCamelCase = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(a , a ) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" pass def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCamelCase = self.rust_tokenizer_class.from_pretrained(a , **a ) __lowerCamelCase = self.tokenizer_class.from_pretrained(a , **a ) __lowerCamelCase = '''A, <mask> AllenNLP sentence.''' __lowerCamelCase = tokenizer_r.encode_plus(a , add_special_tokens=a , return_token_type_ids=a ) __lowerCamelCase = tokenizer_p.encode_plus(a , add_special_tokens=a , return_token_type_ids=a ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r['''token_type_ids'''] ) , sum(tokens_p['''token_type_ids'''] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r['''attention_mask'''] ) / len(tokens_r['''attention_mask'''] ) , sum(tokens_p['''attention_mask'''] ) / len(tokens_p['''attention_mask'''] ) , ) __lowerCamelCase = tokenizer_r.convert_ids_to_tokens(tokens_r['''input_ids'''] ) __lowerCamelCase = tokenizer_p.convert_ids_to_tokens(tokens_p['''input_ids'''] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p['''input_ids'''] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r['''input_ids'''] , [0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( a , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) self.assertSequenceEqual( a , ['''<s>''', '''A''', ''',''', '''<mask>''', '''ĠAllen''', '''N''', '''LP''', '''Ġsentence''', '''.''', '''</s>'''] ) def SCREAMING_SNAKE_CASE__ ( self : str ): """simple docstring""" for trim_offsets, add_prefix_space in itertools.product([True, False] , repeat=2 ): __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( self.tmpdirname , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) __lowerCamelCase = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state['''add_prefix_space'''] , a ) self.assertEqual(post_processor_state['''add_prefix_space'''] , a ) self.assertEqual(post_processor_state['''trim_offsets'''] , a ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __lowerCamelCase = '''hello''' # `hello` is a token in the vocabulary of `pretrained_name` __lowerCamelCase = f"""{text_of_1_token} {text_of_1_token}""" __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( a , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = tokenizer_r(a , return_offsets_mapping=a , add_special_tokens=a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(a ) + 1, len(a ) + 1 + len(a )) , ) __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( a , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = tokenizer_r(a , return_offsets_mapping=a , add_special_tokens=a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(a ) + 1, len(a ) + 1 + len(a )) , ) __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( a , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = tokenizer_r(a , return_offsets_mapping=a , add_special_tokens=a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(a ), len(a ) + 1 + len(a )) , ) __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( a , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = tokenizer_r(a , return_offsets_mapping=a , add_special_tokens=a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(a ), len(a ) + 1 + len(a )) , ) __lowerCamelCase = f""" {text}""" # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( a , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = tokenizer_r(a , return_offsets_mapping=a , add_special_tokens=a ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(a )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(a ) + 1, 1 + len(a ) + 1 + len(a )) , ) __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( a , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = tokenizer_r(a , return_offsets_mapping=a , add_special_tokens=a ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(a )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(a ), 1 + len(a ) + 1 + len(a )) , ) __lowerCamelCase = self.rust_tokenizer_class.from_pretrained( a , use_fast=a , add_prefix_space=a , trim_offsets=a ) __lowerCamelCase = tokenizer_r(a , return_offsets_mapping=a , add_special_tokens=a ) self.assertEqual(encoding.offset_mapping[0] , (0, 1 + len(a )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(a ), 1 + len(a ) + 1 + len(a )) , )
237
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCAmelCase : List[Any] = {'configuration_gpt_neox': ['GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTNeoXConfig']} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : List[Any] = ['GPTNeoXTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : Optional[int] = [ 'GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'GPTNeoXForCausalLM', 'GPTNeoXForQuestionAnswering', 'GPTNeoXForSequenceClassification', 'GPTNeoXForTokenClassification', 'GPTNeoXLayer', 'GPTNeoXModel', 'GPTNeoXPreTrainedModel', ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys __lowerCAmelCase : Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
88
def __lowercase ( _SCREAMING_SNAKE_CASE = 10 ) -> str: '''simple docstring''' if not isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) or n < 0: raise ValueError("""Invalid input""" ) SCREAMING_SNAKE_CASE = 10**n SCREAMING_SNAKE_CASE = 2_84_33 * (pow(2 , 7_83_04_57 , _SCREAMING_SNAKE_CASE )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F'''{solution(1_0) = }''')
296
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Tuple = { """microsoft/beit-base-patch16-224-pt22k""": ( """https://huggingface.co/microsoft/beit-base-patch16-224-pt22k/resolve/main/config.json""" ), # See all BEiT models at https://huggingface.co/models?filter=beit } class _UpperCAmelCase ( __snake_case ): '''simple docstring''' lowerCamelCase__ ='beit' def __init__(self , a_=81_92 , a_=7_68 , a_=12 , a_=12 , a_=30_72 , a_="gelu" , a_=0.0 , a_=0.0 , a_=0.02 , a_=1E-12 , a_=2_24 , a_=16 , a_=3 , a_=False , a_=False , a_=False , a_=False , a_=0.1 , a_=0.1 , a_=True , a_=[3, 5, 7, 11] , a_=[1, 2, 3, 6] , a_=True , a_=0.4 , a_=2_56 , a_=1 , a_=False , a_=2_55 , **a_ , ): '''simple docstring''' super().__init__(**a_ ) __snake_case : str = vocab_size __snake_case : int = hidden_size __snake_case : Optional[Any] = num_hidden_layers __snake_case : List[Any] = num_attention_heads __snake_case : Any = intermediate_size __snake_case : List[Any] = hidden_act __snake_case : Dict = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Any = initializer_range __snake_case : Optional[int] = layer_norm_eps __snake_case : List[Any] = image_size __snake_case : Union[str, Any] = patch_size __snake_case : Optional[int] = num_channels __snake_case : Tuple = use_mask_token __snake_case : int = use_absolute_position_embeddings __snake_case : Any = use_relative_position_bias __snake_case : Any = use_shared_relative_position_bias __snake_case : Tuple = layer_scale_init_value __snake_case : Dict = drop_path_rate __snake_case : Any = use_mean_pooling # decode head attributes (semantic segmentation) __snake_case : Optional[int] = out_indices __snake_case : Optional[int] = pool_scales # auxiliary head attributes (semantic segmentation) __snake_case : str = use_auxiliary_head __snake_case : List[Any] = auxiliary_loss_weight __snake_case : Tuple = auxiliary_channels __snake_case : int = auxiliary_num_convs __snake_case : Dict = auxiliary_concat_input __snake_case : List[Any] = semantic_loss_ignore_index class _UpperCAmelCase ( __snake_case ): '''simple docstring''' lowerCamelCase__ =version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return 1E-4
24
"""simple docstring""" import unittest from transformers import LiltConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, ) from transformers.models.lilt.modeling_lilt import LILT_PRETRAINED_MODEL_ARCHIVE_LIST class _UpperCAmelCase : '''simple docstring''' def __init__(self , a_ , a_=13 , a_=7 , a_=True , a_=True , a_=True , a_=True , a_=99 , a_=24 , a_=2 , a_=6 , a_=37 , a_="gelu" , a_=0.1 , a_=0.1 , a_=5_12 , a_=16 , a_=2 , a_=0.02 , a_=3 , a_=None , a_=10_00 , ): '''simple docstring''' __snake_case : Any = parent __snake_case : int = batch_size __snake_case : Dict = seq_length __snake_case : List[str] = is_training __snake_case : List[Any] = use_input_mask __snake_case : int = use_token_type_ids __snake_case : Union[str, Any] = use_labels __snake_case : str = vocab_size __snake_case : int = hidden_size __snake_case : Optional[int] = num_hidden_layers __snake_case : int = num_attention_heads __snake_case : str = intermediate_size __snake_case : Union[str, Any] = hidden_act __snake_case : int = hidden_dropout_prob __snake_case : Union[str, Any] = attention_probs_dropout_prob __snake_case : List[Any] = max_position_embeddings __snake_case : Any = type_vocab_size __snake_case : Dict = type_sequence_label_size __snake_case : Optional[Any] = initializer_range __snake_case : Union[str, Any] = num_labels __snake_case : Any = scope __snake_case : Any = range_bbox def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : int = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: __snake_case : List[str] = bbox[i, j, 3] __snake_case : Any = bbox[i, j, 1] __snake_case : Tuple = t if bbox[i, j, 2] < bbox[i, j, 0]: __snake_case : List[str] = bbox[i, j, 2] __snake_case : Union[str, Any] = bbox[i, j, 0] __snake_case : Dict = t __snake_case : Optional[int] = None if self.use_input_mask: __snake_case : List[Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) __snake_case : Dict = None if self.use_token_type_ids: __snake_case : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __snake_case : List[str] = None __snake_case : Union[str, Any] = None if self.use_labels: __snake_case : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case : List[Any] = self.get_config() return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return LiltConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , ): '''simple docstring''' __snake_case : Union[str, Any] = LiltModel(config=a_ ) model.to(a_ ) model.eval() __snake_case : Any = model(a_ , bbox=a_ , attention_mask=a_ , token_type_ids=a_ ) __snake_case : str = model(a_ , bbox=a_ , token_type_ids=a_ ) __snake_case : List[str] = model(a_ , bbox=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , ): '''simple docstring''' __snake_case : Optional[int] = self.num_labels __snake_case : List[str] = LiltForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case : Tuple = model( a_ , bbox=a_ , attention_mask=a_ , token_type_ids=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ , a_ , a_ , a_ , a_ , ): '''simple docstring''' __snake_case : Optional[Any] = LiltForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case : int = model( a_ , bbox=a_ , attention_mask=a_ , token_type_ids=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) : Dict = config_and_inputs __snake_case : Any = { '''input_ids''': input_ids, '''bbox''': bbox, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_torch class _UpperCAmelCase ( __snake_case, __snake_case, __snake_case, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =( ( LiltModel, LiltForSequenceClassification, LiltForTokenClassification, LiltForQuestionAnswering, ) if is_torch_available() else () ) lowerCamelCase__ =( { 'feature-extraction': LiltModel, 'question-answering': LiltForQuestionAnswering, 'text-classification': LiltForSequenceClassification, 'token-classification': LiltForTokenClassification, 'zero-shot': LiltForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase__ =False lowerCamelCase__ =False def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ , a_ , a_ ): '''simple docstring''' return True def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = LiltModelTester(self ) __snake_case : Optional[Any] = ConfigTester(self , config_class=a_ , hidden_size=37 ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __snake_case : Dict = type self.model_tester.create_and_check_model(*a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*a_ ) @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' for model_name in LILT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Any = LiltModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) @require_torch @slow class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = LiltModel.from_pretrained('''SCUT-DLVCLab/lilt-roberta-en-base''' ).to(a_ ) __snake_case : Dict = torch.tensor([[1, 2]] , device=a_ ) __snake_case : str = torch.tensor([[[1, 2, 3, 4], [5, 6, 7, 8]]] , device=a_ ) # forward pass with torch.no_grad(): __snake_case : Union[str, Any] = model(input_ids=a_ , bbox=a_ ) __snake_case : Union[str, Any] = torch.Size([1, 2, 7_68] ) __snake_case : str = torch.tensor( [[-0.0653, 0.0950, -0.0061], [-0.0545, 0.0926, -0.0324]] , device=a_ , ) self.assertTrue(outputs.last_hidden_state.shape , a_ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :, :3] , a_ , atol=1E-3 ) )
24
1
from __future__ import annotations __A = "#" class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self : str) ->None: '''simple docstring''' lowerCamelCase__: dict ={} def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : str) ->None: '''simple docstring''' lowerCamelCase__: List[Any] =self._trie for char in text: if char not in trie: lowerCamelCase__: Optional[int] ={} lowerCamelCase__: str =trie[char] lowerCamelCase__: Optional[int] =True def SCREAMING_SNAKE_CASE_ (self : List[Any] , UpperCAmelCase_ : str) ->tuple | list: '''simple docstring''' lowerCamelCase__: Tuple =self._trie for char in prefix: if char in trie: lowerCamelCase__: int =trie[char] else: return [] return self._elements(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : dict) ->tuple: '''simple docstring''' lowerCamelCase__: List[Any] =[] for c, v in d.items(): lowerCamelCase__: int =[" "] if c == END else [(c + s) for s in self._elements(UpperCAmelCase_)] result.extend(UpperCAmelCase_) return tuple(UpperCAmelCase_) __A = Trie() __A = ("depart", "detergent", "daring", "dog", "deer", "deal") for word in words: trie.insert_word(word) def lowerCAmelCase_ ( __a ) -> tuple: """simple docstring""" lowerCamelCase__: Tuple =trie.find_word(__a ) return tuple(string + word for word in suffixes ) def lowerCAmelCase_ ( ) -> None: """simple docstring""" print(autocomplete_using_trie("de" ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
10
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __A = { "configuration_distilbert": [ "DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DistilBertConfig", "DistilBertOnnxConfig", ], "tokenization_distilbert": ["DistilBertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = ["DistilBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "DistilBertForMaskedLM", "DistilBertForMultipleChoice", "DistilBertForQuestionAnswering", "DistilBertForSequenceClassification", "DistilBertForTokenClassification", "DistilBertModel", "DistilBertPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDistilBertForMaskedLM", "TFDistilBertForMultipleChoice", "TFDistilBertForQuestionAnswering", "TFDistilBertForSequenceClassification", "TFDistilBertForTokenClassification", "TFDistilBertMainLayer", "TFDistilBertModel", "TFDistilBertPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A = [ "FlaxDistilBertForMaskedLM", "FlaxDistilBertForMultipleChoice", "FlaxDistilBertForQuestionAnswering", "FlaxDistilBertForSequenceClassification", "FlaxDistilBertForTokenClassification", "FlaxDistilBertModel", "FlaxDistilBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys __A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
10
1
import datasets from .evaluate import evaluate _lowercase : List[Any] = '''\ @inproceedings{Rajpurkar2016SQuAD10, title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text}, author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang}, booktitle={EMNLP}, year={2016} } ''' _lowercase : Optional[Any] = ''' This metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD). Stanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span, from the corresponding reading passage, or the question might be unanswerable. ''' _lowercase : Any = ''' Computes SQuAD scores (F1 and EM). Args: predictions: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair as given in the references (see below) - \'prediction_text\': the text of the answer references: List of question-answers dictionaries with the following key-values: - \'id\': id of the question-answer pair (see above), - \'answers\': a Dict in the SQuAD dataset format { \'text\': list of possible texts for the answer, as a list of strings \'answer_start\': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: \'exact_match\': Exact match (the normalized answer exactly match the gold answer) \'f1\': The F-score of predicted tokens versus the gold answer Examples: >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}] >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}] >>> squad_metric = datasets.load_metric("squad") >>> results = squad_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 100.0, \'f1\': 100.0} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase__ ( datasets.Metric ): def _snake_case ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': {'''id''': datasets.Value('''string''' ), '''prediction_text''': datasets.Value('''string''' )}, '''references''': { '''id''': datasets.Value('''string''' ), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string''' ), '''answer_start''': datasets.Value('''int32''' ), } ), }, } ) , codebase_urls=['''https://rajpurkar.github.io/SQuAD-explorer/'''] , reference_urls=['''https://rajpurkar.github.io/SQuAD-explorer/'''] , ) def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Tuple = {prediction['id']: prediction['prediction_text'] for prediction in predictions} lowercase_ : Union[str, Any] = [ { 'paragraphs': [ { 'qas': [ { 'answers': [{'text': answer_text} for answer_text in ref['answers']['text']], 'id': ref['id'], } for ref in references ] } ] } ] lowercase_ : Dict = evaluate(dataset=__a , predictions=__a ) return score
350
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class lowerCAmelCase__ ( lowerCamelCase_ ): lowerCAmelCase_ = (DDPMScheduler,) def _snake_case ( self , **__SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Optional[Any] = { '''num_train_timesteps''': 10_00, '''beta_start''': 0.0_001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''variance_type''': '''fixed_small''', '''clip_sample''': True, } config.update(**__SCREAMING_SNAKE_CASE ) return config def _snake_case ( self ): """simple docstring""" for timesteps in [1, 5, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" for beta_start, beta_end in zip([0.0_001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=__SCREAMING_SNAKE_CASE , beta_end=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" for clip_sample in [True, False]: self.check_over_configs(clip_sample=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" self.check_over_configs(thresholding=__SCREAMING_SNAKE_CASE ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=__SCREAMING_SNAKE_CASE , prediction_type=__SCREAMING_SNAKE_CASE , sample_max_value=__SCREAMING_SNAKE_CASE , ) def _snake_case ( self ): """simple docstring""" for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" for t in [0, 5_00, 9_99]: self.check_over_forward(time_step=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ : List[str] = self.scheduler_classes[0] lowercase_ : int = self.get_scheduler_config() lowercase_ : str = scheduler_class(**__SCREAMING_SNAKE_CASE ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_87 ) - 0.00_979 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_99 ) - 0.02 ) ) < 1E-5 def _snake_case ( self ): """simple docstring""" lowercase_ : Union[str, Any] = self.scheduler_classes[0] lowercase_ : Any = self.get_scheduler_config() lowercase_ : Optional[int] = scheduler_class(**__SCREAMING_SNAKE_CASE ) lowercase_ : Any = len(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = self.dummy_model() lowercase_ : Any = self.dummy_sample_deter lowercase_ : List[str] = torch.manual_seed(0 ) for t in reversed(range(__SCREAMING_SNAKE_CASE ) ): # 1. predict noise residual lowercase_ : Optional[int] = model(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # 2. predict previous mean of sample x_t-1 lowercase_ : Union[str, Any] = scheduler.step(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance lowercase_ : List[str] = pred_prev_sample lowercase_ : str = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE ) ) lowercase_ : Tuple = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 258.9_606 ) < 1E-2 assert abs(result_mean.item() - 0.3_372 ) < 1E-3 def _snake_case ( self ): """simple docstring""" lowercase_ : Optional[Any] = self.scheduler_classes[0] lowercase_ : Tuple = self.get_scheduler_config(prediction_type='''v_prediction''' ) lowercase_ : Any = scheduler_class(**__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = len(__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = self.dummy_model() lowercase_ : int = self.dummy_sample_deter lowercase_ : str = torch.manual_seed(0 ) for t in reversed(range(__SCREAMING_SNAKE_CASE ) ): # 1. predict noise residual lowercase_ : Optional[Any] = model(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # 2. predict previous mean of sample x_t-1 lowercase_ : Optional[Any] = scheduler.step(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance lowercase_ : int = pred_prev_sample lowercase_ : str = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE ) ) lowercase_ : Tuple = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 202.0_296 ) < 1E-2 assert abs(result_mean.item() - 0.2_631 ) < 1E-3 def _snake_case ( self ): """simple docstring""" lowercase_ : Optional[int] = self.scheduler_classes[0] lowercase_ : int = self.get_scheduler_config() lowercase_ : int = scheduler_class(**__SCREAMING_SNAKE_CASE ) lowercase_ : int = [1_00, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = scheduler.timesteps for i, timestep in enumerate(__SCREAMING_SNAKE_CASE ): if i == len(__SCREAMING_SNAKE_CASE ) - 1: lowercase_ : str = -1 else: lowercase_ : Any = timesteps[i + 1] lowercase_ : Optional[Any] = scheduler.previous_timestep(__SCREAMING_SNAKE_CASE ) lowercase_ : int = prev_t.item() self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ : Tuple = self.scheduler_classes[0] lowercase_ : List[Any] = self.get_scheduler_config() lowercase_ : str = scheduler_class(**__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = [1_00, 87, 50, 51, 0] with self.assertRaises(__SCREAMING_SNAKE_CASE , msg='''`custom_timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ : Optional[Any] = self.scheduler_classes[0] lowercase_ : str = self.get_scheduler_config() lowercase_ : Tuple = scheduler_class(**__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = [1_00, 87, 50, 1, 0] lowercase_ : List[str] = len(__SCREAMING_SNAKE_CASE ) with self.assertRaises(__SCREAMING_SNAKE_CASE , msg='''Can only pass one of `num_inference_steps` or `custom_timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=__SCREAMING_SNAKE_CASE , timesteps=__SCREAMING_SNAKE_CASE ) def _snake_case ( self ): """simple docstring""" lowercase_ : List[Any] = self.scheduler_classes[0] lowercase_ : Optional[int] = self.get_scheduler_config() lowercase_ : Any = scheduler_class(**__SCREAMING_SNAKE_CASE ) lowercase_ : str = [scheduler.config.num_train_timesteps] with self.assertRaises( __SCREAMING_SNAKE_CASE , msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' , ): scheduler.set_timesteps(timesteps=__SCREAMING_SNAKE_CASE )
264
0
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class _A : def __init__( self : List[str] , _A : str = "cpu" , _A : str = "openai/clip-vit-large-patch14" ) -> str: """simple docstring""" lowercase : Optional[int] = device lowercase : Any = CLIPTokenizerFast.from_pretrained(a__ ) lowercase : str = [0.48_145_466, 0.4_578_275, 0.40_821_073] lowercase : Optional[int] = [0.26_862_954, 0.26_130_258, 0.27_577_711] lowercase : Optional[Any] = torchvision.transforms.Normalize(self.image_mean , self.image_std ) lowercase : Union[str, Any] = torchvision.transforms.Resize(224 ) lowercase : List[Any] = torchvision.transforms.CenterCrop(224 ) def __a ( self : Optional[Any] , _A : str ) -> Union[str, Any]: """simple docstring""" lowercase : Dict = self.resize(a__ ) lowercase : Optional[Any] = self.center_crop(a__ ) lowercase : List[str] = self.normalize(a__ ) return images def __call__( self : Union[str, Any] , _A : Tuple=None , _A : Tuple=None , **_A : int ) -> Union[str, Any]: """simple docstring""" lowercase : Optional[int] = self.tokenizer(text=a__ , **a__ ) lowercase : Any = self.preprocess_img(a__ ) lowercase : Optional[Any] = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class _A ( nn.Module ): def __init__( self : Optional[int] , _A : Dict=10 , _A : Dict=0.01 , _A : Union[str, Any]=None , _A : Any=None , _A : Optional[Any]=None , _A : Dict=None , _A : Union[str, Any]=None , _A : List[str]=None , _A : Union[str, Any]=False , _A : Dict=True , _A : Dict="image" , _A : Optional[Any]=True , _A : List[str]=False , _A : Optional[Any]=False , _A : List[str]=False , ) -> Any: """simple docstring""" super().__init__() lowercase : Optional[Any] = None lowercase : List[Any] = device if device else get_device() if vqgan: lowercase : Tuple = vqgan else: lowercase : int = load_vqgan(self.device , conf_path=a__ , ckpt_path=a__ ) self.vqgan.eval() if clip: lowercase : Any = clip else: lowercase : List[str] = CLIPModel.from_pretrained('''openai/clip-vit-base-patch32''' ) self.clip.to(self.device ) lowercase : Tuple = ProcessorGradientFlow(device=self.device ) lowercase : str = iterations lowercase : Optional[int] = lr lowercase : Optional[Any] = log lowercase : Union[str, Any] = make_grid lowercase : List[Any] = return_val lowercase : List[str] = quantize lowercase : int = self.vqgan.decoder.z_shape def __a ( self : str , _A : Optional[int]=None , _A : Tuple=None , _A : str=5 , _A : int=True ) -> str: """simple docstring""" lowercase : Any = [] if output_path is None: lowercase : int = '''./animation.gif''' if input_path is None: lowercase : List[Any] = self.save_path lowercase : Optional[int] = sorted(glob(input_path + '''/*''' ) ) if not len(a__ ): raise ValueError( '''No images found in save path, aborting (did you pass save_intermediate=True to the generate''' ''' function?)''' ) if len(a__ ) == 1: print('''Only one image found in save path, (did you pass save_intermediate=True to the generate function?)''' ) lowercase : List[Any] = total_duration / len(a__ ) lowercase : int = [frame_duration] * len(a__ ) if extend_frames: lowercase : Dict = 1.5 lowercase : Any = 3 for file_name in paths: if file_name.endswith('''.png''' ): images.append(imageio.imread(a__ ) ) imageio.mimsave(a__ , a__ , duration=a__ ) print(f"""gif saved to {output_path}""" ) def __a ( self : List[Any] , _A : List[str]=None , _A : Optional[int]=None ) -> Tuple: """simple docstring""" if not (path or img): raise ValueError('''Input either path or tensor''' ) if img is not None: raise NotImplementedError lowercase : int = preprocess(Image.open(a__ ) , target_image_size=256 ).to(self.device ) lowercase : Union[str, Any] = preprocess_vqgan(a__ ) lowercase , *lowercase : Tuple = self.vqgan.encode(a__ ) return z def __a ( self : List[str] , _A : int ) -> Dict: """simple docstring""" lowercase : List[Any] = self.latent.detach().requires_grad_() lowercase : List[str] = base_latent + transform_vector if self.quantize: lowercase , *lowercase : Optional[int] = self.vqgan.quantize(a__ ) else: lowercase : int = trans_latent return self.vqgan.decode(a__ ) def __a ( self : int , _A : Dict , _A : str , _A : str=None ) -> Tuple: """simple docstring""" lowercase : Any = self.clip_preprocessor(text=a__ , images=a__ , return_tensors='''pt''' , padding=a__ ) lowercase : Dict = self.clip(**a__ ) lowercase : Dict = clip_outputs.logits_per_image if weights is not None: lowercase : Dict = similarity_logits * weights return similarity_logits.sum() def __a ( self : Dict , _A : Union[str, Any] , _A : Union[str, Any] , _A : int ) -> List[Any]: """simple docstring""" lowercase : Optional[int] = self._get_clip_similarity(pos_prompts['''prompts'''] , a__ , weights=(1 / pos_prompts['''weights''']) ) if neg_prompts: lowercase : Any = self._get_clip_similarity(neg_prompts['''prompts'''] , a__ , weights=neg_prompts['''weights'''] ) else: lowercase : Any = torch.tensor([1] , device=self.device ) lowercase : Dict = -torch.log(a__ ) + torch.log(a__ ) return loss def __a ( self : Any , _A : Optional[Any] , _A : List[str] , _A : Tuple ) -> List[Any]: """simple docstring""" lowercase : int = torch.randn_like(self.latent , requires_grad=a__ , device=self.device ) lowercase : int = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() lowercase : List[Any] = self._add_vector(a__ ) lowercase : Tuple = loop_post_process(a__ ) lowercase : Any = self._get_CLIP_loss(a__ , a__ , a__ ) print('''CLIP loss''' , a__ ) if self.log: wandb.log({'''CLIP Loss''': clip_loss} ) clip_loss.backward(retain_graph=a__ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def __a ( self : int , _A : Optional[Any] , _A : Optional[int] , _A : List[Any] ) -> Any: """simple docstring""" wandb.init(reinit=a__ , project='''face-editor''' ) wandb.config.update({'''Positive Prompts''': positive_prompts} ) wandb.config.update({'''Negative Prompts''': negative_prompts} ) wandb.config.update({'''lr''': self.lr, '''iterations''': self.iterations} ) if image_path: lowercase : Optional[int] = Image.open(a__ ) lowercase : Any = image.resize((256, 256) ) wandb.log('''Original Image''' , wandb.Image(a__ ) ) def __a ( self : Optional[int] , _A : int ) -> List[str]: """simple docstring""" if not prompts: return [] lowercase : str = [] lowercase : Optional[int] = [] if isinstance(a__ , a__ ): lowercase : Optional[int] = [prompt.strip() for prompt in prompts.split('''|''' )] for prompt in prompts: if isinstance(a__ , (tuple, list) ): lowercase : Any = prompt[0] lowercase : List[Any] = float(prompt[1] ) elif ":" in prompt: lowercase , lowercase : List[Any] = prompt.split(''':''' ) lowercase : int = float(a__ ) else: lowercase : int = prompt lowercase : int = 1.0 processed_prompts.append(a__ ) weights.append(a__ ) return { "prompts": processed_prompts, "weights": torch.tensor(a__ , device=self.device ), } def __a ( self : Optional[Any] , _A : Union[str, Any] , _A : Any=None , _A : List[Any]=None , _A : Optional[int]=True , _A : Any=False , _A : Any=True , _A : int=True , _A : List[str]=None , ) -> Any: """simple docstring""" if image_path: lowercase : List[str] = self._get_latent(a__ ) else: lowercase : Union[str, Any] = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(a__ , a__ , a__ ) assert pos_prompts, "You must provide at least one positive prompt." lowercase : str = self.process_prompts(a__ ) lowercase : int = self.process_prompts(a__ ) if save_final and save_path is None: lowercase : List[Any] = os.path.join('''./outputs/''' , '''_'''.join(pos_prompts['''prompts'''] ) ) if not os.path.exists(a__ ): os.makedirs(a__ ) else: lowercase : Optional[Any] = save_path + '''_''' + get_timestamp() os.makedirs(a__ ) lowercase : Tuple = save_path lowercase : Tuple = self.vqgan.decode(self.latent )[0] if show_intermediate: print('''Original Image''' ) show_pil(custom_to_pil(a__ ) ) lowercase : Optional[Any] = loop_post_process(a__ ) for iter, transformed_img in enumerate(self._optimize_CLIP(a__ , a__ , a__ ) ): if show_intermediate: show_pil(a__ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , f"""iter_{iter:03d}.png""" ) ) if self.log: wandb.log({'''Image''': wandb.Image(a__ )} ) if show_final: show_pil(a__ ) if save_final: transformed_img.save(os.path.join(self.save_path , f"""iter_{iter:03d}_final.png""" ) )
308
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin snake_case_ = get_tests_dir('fixtures/test_sentencepiece_bpe.model') class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , unittest.TestCase ): A_ : List[Any] = BartphoTokenizer A_ : List[str] = False A_ : Optional[Any] = True def a (self : Tuple ): """simple docstring""" super().setUp() __snake_case = ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] __snake_case = dict(zip(a__ , range(len(a__ ) ) ) ) __snake_case = {'''unk_token''': '''<unk>'''} __snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''monolingual_vocab_file'''] ) with open(self.monolingual_vocab_file , '''w''' , encoding='''utf-8''' ) as fp: for token in vocab_tokens: fp.write(f"""{token} {vocab_tokens[token]}\n""" ) __snake_case = BartphoTokenizer(a__ , self.monolingual_vocab_file , **self.special_tokens_map ) tokenizer.save_pretrained(self.tmpdirname ) def a (self : str , **a__ : str ): """simple docstring""" kwargs.update(self.special_tokens_map ) return BartphoTokenizer.from_pretrained(self.tmpdirname , **a__ ) def a (self : str , a__ : Any ): """simple docstring""" __snake_case = '''This is a là test''' __snake_case = '''This is a<unk><unk> test''' return input_text, output_text def a (self : Dict ): """simple docstring""" __snake_case = BartphoTokenizer(a__ , self.monolingual_vocab_file , **self.special_tokens_map ) __snake_case = '''This is a là test''' __snake_case = '''▁This ▁is ▁a ▁l à ▁t est'''.split() __snake_case = tokenizer.tokenize(a__ ) self.assertListEqual(a__ , a__ ) __snake_case = tokens + [tokenizer.unk_token] __snake_case = [4, 5, 6, 3, 3, 7, 8, 3] self.assertListEqual(tokenizer.convert_tokens_to_ids(a__ ) , a__ )
24
0
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor lowerCAmelCase_ = logging.get_logger(__name__) class _lowerCAmelCase ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self : str , *UpperCamelCase : Optional[Any] , **UpperCamelCase : Optional[int] ): '''simple docstring''' warnings.warn( 'The class BeitFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use BeitImageProcessor instead.' , UpperCamelCase , ) super().__init__(*UpperCamelCase , **UpperCamelCase )
260
from __future__ import annotations lowerCAmelCase_ = [] def lowerCamelCase_ ( lowerCAmelCase: list[list[int]] , lowerCAmelCase: int , lowerCAmelCase: int )-> bool: for i in range(len(lowerCAmelCase ) ): if board[row][i] == 1: return False for i in range(len(lowerCAmelCase ) ): if board[i][column] == 1: return False for i, j in zip(range(lowerCAmelCase , -1 , -1 ) , range(lowerCAmelCase , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(lowerCAmelCase , -1 , -1 ) , range(lowerCAmelCase , len(lowerCAmelCase ) ) ): if board[i][j] == 1: return False return True def lowerCamelCase_ ( lowerCAmelCase: list[list[int]] , lowerCAmelCase: int )-> bool: if row >= len(lowerCAmelCase ): solution.append(lowerCAmelCase ) printboard(lowerCAmelCase ) print() return True for i in range(len(lowerCAmelCase ) ): if is_safe(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): _snake_case : Dict = 1 solve(lowerCAmelCase , row + 1 ) _snake_case : str = 0 return False def lowerCamelCase_ ( lowerCAmelCase: list[list[int]] )-> None: for i in range(len(lowerCAmelCase ) ): for j in range(len(lowerCAmelCase ) ): if board[i][j] == 1: print('Q' , end=' ' ) else: print('.' , end=' ' ) print() # n=int(input("The no. of queens")) lowerCAmelCase_ = 8 lowerCAmelCase_ = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print("""The total no. of solutions are :""", len(solution))
260
1
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 snake_case_ : List[Any] = data_utils.TransfoXLTokenizer snake_case_ : int = data_utils.TransfoXLCorpus snake_case_ : List[Any] = data_utils snake_case_ : int = data_utils def A (__A : Dict , __A : List[Any] , __A : Union[str, Any] , __A : Tuple ) -> Union[str, Any]: """simple docstring""" if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(__A , '''rb''' ) as fp: UpperCAmelCase_ = pickle.load(__A , encoding='''latin1''' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) UpperCAmelCase_ = pytorch_dump_folder_path + '''/''' + VOCAB_FILES_NAMES['''pretrained_vocab_file'''] print(F"""Save vocabulary to {pytorch_vocab_dump_path}""" ) UpperCAmelCase_ = corpus.vocab.__dict__ torch.save(__A , __A ) UpperCAmelCase_ = corpus.__dict__ corpus_dict_no_vocab.pop('''vocab''' , __A ) UpperCAmelCase_ = pytorch_dump_folder_path + '''/''' + CORPUS_NAME print(F"""Save dataset to {pytorch_dataset_dump_path}""" ) torch.save(__A , __A ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model UpperCAmelCase_ = os.path.abspath(__A ) UpperCAmelCase_ = os.path.abspath(__A ) print(F"""Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.""" ) # Initialise PyTorch model if transfo_xl_config_file == "": UpperCAmelCase_ = TransfoXLConfig() else: UpperCAmelCase_ = TransfoXLConfig.from_json_file(__A ) print(F"""Building PyTorch model from configuration: {config}""" ) UpperCAmelCase_ = TransfoXLLMHeadModel(__A ) UpperCAmelCase_ = load_tf_weights_in_transfo_xl(__A , __A , __A ) # Save pytorch-model UpperCAmelCase_ = os.path.join(__A , __A ) UpperCAmelCase_ = os.path.join(__A , __A ) print(F"""Save PyTorch model to {os.path.abspath(__A )}""" ) torch.save(model.state_dict() , __A ) print(F"""Save configuration file to {os.path.abspath(__A )}""" ) with open(__A , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": snake_case_ : List[str] = argparse.ArgumentParser() parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the folder to store the PyTorch model or dataset/vocab.", ) parser.add_argument( "--tf_checkpoint_path", default="", type=str, help="An optional path to a TensorFlow checkpoint path to be converted.", ) parser.add_argument( "--transfo_xl_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained BERT model. \n" "This specifies the model architecture." ), ) parser.add_argument( "--transfo_xl_dataset_file", default="", type=str, help="An optional dataset file to be converted in a vocabulary.", ) snake_case_ : int = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
51
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case : List[Any] = logging.get_logger(__name__) _snake_case : List[Any] = "▁" _snake_case : Union[str, Any] = {"vocab_file": "sentencepiece.bpe.model"} _snake_case : Any = { "vocab_file": { "facebook/xglm-564M": "https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model", } } _snake_case : Union[str, Any] = { "facebook/xglm-564M": 2_048, } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Dict = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Any = ["input_ids", "attention_mask"] def __init__( self : Tuple , lowerCamelCase : Tuple , lowerCamelCase : Optional[int]="<s>" , lowerCamelCase : int="</s>" , lowerCamelCase : Dict="</s>" , lowerCamelCase : Tuple="<s>" , lowerCamelCase : List[str]="<unk>" , lowerCamelCase : str="<pad>" , lowerCamelCase : Optional[Dict[str, Any]] = None , **lowerCamelCase : int , ) -> None: __snake_case : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer __snake_case : Tuple = 7 __snake_case : Optional[int] = [F'<madeupword{i}>' for i in range(self.num_madeup_words )] __snake_case : Tuple = kwargs.get("additional_special_tokens" , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=lowerCamelCase , eos_token=lowerCamelCase , unk_token=lowerCamelCase , sep_token=lowerCamelCase , cls_token=lowerCamelCase , pad_token=lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase , ) __snake_case : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCamelCase ) ) __snake_case : List[str] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __snake_case : Optional[Any] = 1 # Mimic fairseq token-to-id alignment for the first 4 token __snake_case : List[Any] = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} __snake_case : Union[str, Any] = len(self.sp_model ) __snake_case : Union[str, Any] = {F'<madeupword{i}>': sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(lowerCamelCase ) __snake_case : Dict = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Dict ) -> List[Any]: __snake_case : Any = self.__dict__.copy() __snake_case : str = None __snake_case : str = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple , lowerCamelCase : Union[str, Any] ) -> Optional[Any]: __snake_case : Optional[Any] = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __snake_case : List[str] = {} __snake_case : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def __snake_case ( self : List[str] , lowerCamelCase : List[int] , lowerCamelCase : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.sep_token_id] + token_ids_a __snake_case : int = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def __snake_case ( self : Dict , lowerCamelCase : List[int] , lowerCamelCase : Optional[List[int]] = None , lowerCamelCase : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase , token_ids_a=lowerCamelCase , already_has_special_tokens=lowerCamelCase ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase )) return [1] + ([0] * len(lowerCamelCase )) + [1, 1] + ([0] * len(lowerCamelCase )) def __snake_case ( self : Dict , lowerCamelCase : List[int] , lowerCamelCase : Optional[List[int]] = None ) -> List[int]: __snake_case : str = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def __snake_case ( self : Any ) -> List[str]: return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def __snake_case ( self : Union[str, Any] ) -> Tuple: __snake_case : Optional[int] = {self.convert_ids_to_tokens(lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __snake_case ( self : int , lowerCamelCase : str ) -> List[str]: return self.sp_model.encode(lowerCamelCase , out_type=lowerCamelCase ) def __snake_case ( self : Any , lowerCamelCase : Any ) -> Optional[int]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __snake_case : Optional[Any] = self.sp_model.PieceToId(lowerCamelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __snake_case ( self : Any , lowerCamelCase : int ) -> List[str]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __snake_case ( self : int , lowerCamelCase : str ) -> Tuple: __snake_case : Optional[Any] = "".join(lowerCamelCase ).replace(lowerCamelCase , " " ).strip() return out_string def __snake_case ( self : List[str] , lowerCamelCase : str , lowerCamelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCamelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __snake_case : str = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase , "wb" ) as fi: __snake_case : Any = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase ) return (out_vocab_file,)
123
0
from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES _UpperCAmelCase : Any = 'tiny-wmt19-en-ru' # Build # borrowed from a test _UpperCAmelCase : str = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] _UpperCAmelCase : List[Any] = dict(zip(vocab, range(len(vocab)))) _UpperCAmelCase : str = ['l o 123', 'lo w 1456', 'e r</w> 1789', ''] with tempfile.TemporaryDirectory() as tmpdirname: _UpperCAmelCase : int = Path(tmpdirname) _UpperCAmelCase : Union[str, Any] = build_dir / VOCAB_FILES_NAMES['src_vocab_file'] _UpperCAmelCase : Dict = build_dir / VOCAB_FILES_NAMES['tgt_vocab_file'] _UpperCAmelCase : List[str] = build_dir / VOCAB_FILES_NAMES['merges_file'] with open(src_vocab_file, "w") as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, "w") as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, "w") as fp: fp.write("\n".join(merges)) _UpperCAmelCase : Optional[int] = FSMTTokenizer( langs=["en", "ru"], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) _UpperCAmelCase : Union[str, Any] = FSMTConfig( langs=["ru", "en"], src_vocab_size=1000, tgt_vocab_size=1000, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) _UpperCAmelCase : Optional[Any] = FSMTForConditionalGeneration(config) print(f'''num of params {tiny_model.num_parameters()}''') # Test _UpperCAmelCase : List[str] = tokenizer(["Making tiny model"], return_tensors="pt") _UpperCAmelCase : Tuple = tiny_model(**batch) print("test output:", len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(f'''Generated {mname_tiny}''') # Upload # transformers-cli upload tiny-wmt19-en-ru
358
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPanoramaPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() @skip_mps class __lowerCAmelCase ( lowerCAmelCase , lowerCAmelCase , unittest.TestCase): _a = StableDiffusionPanoramaPipeline _a = TEXT_TO_IMAGE_PARAMS _a = TEXT_TO_IMAGE_BATCH_PARAMS _a = TEXT_TO_IMAGE_IMAGE_PARAMS _a = TEXT_TO_IMAGE_IMAGE_PARAMS def SCREAMING_SNAKE_CASE ( self: int ): torch.manual_seed(0 ) lowercase :List[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) lowercase :Any = DDIMScheduler() torch.manual_seed(0 ) lowercase :Any = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) lowercase :Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) lowercase :Any = CLIPTextModel(_lowerCAmelCase ) lowercase :str = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowercase :Union[str, Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def SCREAMING_SNAKE_CASE ( self: int , _lowerCAmelCase: List[Any] , _lowerCAmelCase: Dict=0 ): lowercase :Any = torch.manual_seed(_lowerCAmelCase ) lowercase :Any = { "prompt": "a photo of the dolomites", "generator": generator, # Setting height and width to None to prevent OOMs on CPU. "height": None, "width": None, "num_inference_steps": 1, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE ( self: List[Any] ): lowercase :List[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase :int = self.get_dummy_components() lowercase :int = StableDiffusionPanoramaPipeline(**_lowerCAmelCase ) lowercase :Tuple = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :List[str] = self.get_dummy_inputs(_lowerCAmelCase ) lowercase :List[Any] = sd_pipe(**_lowerCAmelCase ).images lowercase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase :List[str] = np.array([0.61_86, 0.53_74, 0.49_15, 0.41_35, 0.41_14, 0.45_63, 0.51_28, 0.49_77, 0.47_57] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): super().test_inference_batch_consistent(batch_sizes=[1, 2] ) def SCREAMING_SNAKE_CASE ( self: int ): super().test_inference_batch_single_identical(batch_size=2 , expected_max_diff=3.2_5e-3 ) def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): lowercase :Optional[int] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase :List[str] = self.get_dummy_components() lowercase :Optional[Any] = StableDiffusionPanoramaPipeline(**_lowerCAmelCase ) lowercase :Any = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Optional[int] = self.get_dummy_inputs(_lowerCAmelCase ) lowercase :List[Any] = "french fries" lowercase :int = sd_pipe(**_lowerCAmelCase , negative_prompt=_lowerCAmelCase ) lowercase :int = output.images lowercase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase :Optional[Any] = np.array([0.61_87, 0.53_75, 0.49_15, 0.41_36, 0.41_14, 0.45_63, 0.51_28, 0.49_76, 0.47_57] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self: List[str] ): lowercase :Optional[int] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase :int = self.get_dummy_components() lowercase :List[str] = StableDiffusionPanoramaPipeline(**_lowerCAmelCase ) lowercase :int = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Dict = self.get_dummy_inputs(_lowerCAmelCase ) lowercase :Any = sd_pipe(**_lowerCAmelCase , view_batch_size=2 ) lowercase :Union[str, Any] = output.images lowercase :List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase :Optional[int] = np.array([0.61_87, 0.53_75, 0.49_15, 0.41_36, 0.41_14, 0.45_63, 0.51_28, 0.49_76, 0.47_57] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self: str ): lowercase :int = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase :List[Any] = self.get_dummy_components() lowercase :Tuple = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" ) lowercase :Tuple = StableDiffusionPanoramaPipeline(**_lowerCAmelCase ) lowercase :Any = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Optional[Any] = self.get_dummy_inputs(_lowerCAmelCase ) lowercase :List[Any] = sd_pipe(**_lowerCAmelCase ).images lowercase :Any = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase :Optional[Any] = np.array([0.40_24, 0.65_10, 0.49_01, 0.53_78, 0.58_13, 0.56_22, 0.47_95, 0.44_67, 0.49_52] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): lowercase :Optional[int] = "cpu" # ensure determinism for the device-dependent torch.Generator lowercase :List[Any] = self.get_dummy_components() lowercase :int = PNDMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , skip_prk_steps=_lowerCAmelCase ) lowercase :List[str] = StableDiffusionPanoramaPipeline(**_lowerCAmelCase ) lowercase :int = sd_pipe.to(_lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=_lowerCAmelCase ) lowercase :Union[str, Any] = self.get_dummy_inputs(_lowerCAmelCase ) lowercase :Any = sd_pipe(**_lowerCAmelCase ).images lowercase :Any = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowercase :Tuple = np.array([0.63_91, 0.62_91, 0.48_61, 0.51_34, 0.55_52, 0.45_78, 0.50_32, 0.50_23, 0.45_39] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self: str ): super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: Union[str, Any]=0 ): lowercase :Any = torch.manual_seed(_lowerCAmelCase ) lowercase :Optional[int] = { "prompt": "a photo of the dolomites", "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def SCREAMING_SNAKE_CASE ( self: Tuple ): lowercase :Dict = "stabilityai/stable-diffusion-2-base" lowercase :Optional[Any] = DDIMScheduler.from_pretrained(_lowerCAmelCase , subfolder="scheduler" ) lowercase :List[str] = StableDiffusionPanoramaPipeline.from_pretrained(_lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() lowercase :Any = self.get_inputs() lowercase :Optional[int] = pipe(**_lowerCAmelCase ).images lowercase :int = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 20_48, 3) lowercase :Tuple = np.array( [ 0.36_96_83_92, 0.27_02_53_72, 0.32_44_67_66, 0.28_37_93_87, 0.36_36_32_74, 0.30_73_33_47, 0.27_10_00_27, 0.27_05_41_25, 0.25_53_60_96, ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE ( self: Union[str, Any] ): lowercase :Union[str, Any] = StableDiffusionPanoramaPipeline.from_pretrained( "stabilityai/stable-diffusion-2-base" , safety_checker=_lowerCAmelCase ) lowercase :int = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() lowercase :int = self.get_inputs() lowercase :Optional[int] = pipe(**_lowerCAmelCase ).images lowercase :List[str] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 20_48, 3) lowercase :Union[str, Any] = np.array( [ [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, ] ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def SCREAMING_SNAKE_CASE ( self: List[str] ): lowercase :Dict = 0 def callback_fn(_lowerCAmelCase: int , _lowerCAmelCase: int , _lowerCAmelCase: torch.FloatTensor ) -> None: lowercase :Any = True nonlocal number_of_steps number_of_steps += 1 if step == 1: lowercase :Union[str, Any] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 2_56) lowercase :Optional[int] = latents[0, -3:, -3:, -1] lowercase :Any = np.array( [ 0.18_68_18_69, 0.33_90_78_16, 0.5_36_12_76, 0.14_43_28_65, -0.02_85_66_11, -0.73_94_11_23, 0.23_39_79_87, 0.47_32_26_82, -0.37_82_31_64, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: lowercase :str = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 2_56) lowercase :Optional[int] = latents[0, -3:, -3:, -1] lowercase :Optional[Any] = np.array( [ 0.18_53_96_45, 0.33_98_72_48, 0.5_37_85_59, 0.14_43_71_42, -0.02_45_52_61, -0.7_33_83_17, 0.23_99_07_55, 0.47_35_62_72, -0.3_78_65_05, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 lowercase :int = False lowercase :Tuple = "stabilityai/stable-diffusion-2-base" lowercase :Optional[Any] = DDIMScheduler.from_pretrained(_lowerCAmelCase , subfolder="scheduler" ) lowercase :Optional[int] = StableDiffusionPanoramaPipeline.from_pretrained(_lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase ) lowercase :Optional[int] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing() lowercase :int = self.get_inputs() pipe(**_lowerCAmelCase , callback=_lowerCAmelCase , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def SCREAMING_SNAKE_CASE ( self: str ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() lowercase :Optional[Any] = "stabilityai/stable-diffusion-2-base" lowercase :Dict = DDIMScheduler.from_pretrained(_lowerCAmelCase , subfolder="scheduler" ) lowercase :Optional[int] = StableDiffusionPanoramaPipeline.from_pretrained(_lowerCAmelCase , scheduler=_lowerCAmelCase , safety_checker=_lowerCAmelCase ) lowercase :Union[str, Any] = pipe.to(_lowerCAmelCase ) pipe.set_progress_bar_config(disable=_lowerCAmelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() lowercase :Optional[int] = self.get_inputs() lowercase :Union[str, Any] = pipe(**_lowerCAmelCase ) lowercase :List[Any] = torch.cuda.max_memory_allocated() # make sure that less than 5.2 GB is allocated assert mem_bytes < 5.5 * 10**9
158
0
import math def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ = 0 , lowerCamelCase__ = 0 ): lowerCamelCase_ = end or len(lowerCamelCase__ ) for i in range(lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = i lowerCamelCase_ = array[i] while temp_index != start and temp_index_value < array[temp_index - 1]: lowerCamelCase_ = array[temp_index - 1] temp_index -= 1 lowerCamelCase_ = temp_index_value return array def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): # Max Heap lowerCamelCase_ = index lowerCamelCase_ = 2 * index + 1 # Left Node lowerCamelCase_ = 2 * index + 2 # Right Node if left_index < heap_size and array[largest] < array[left_index]: lowerCamelCase_ = left_index if right_index < heap_size and array[largest] < array[right_index]: lowerCamelCase_ = right_index if largest != index: lowerCamelCase_ , lowerCamelCase_ = array[largest], array[index] heapify(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = len(lowerCamelCase__ ) for i in range(n // 2 , -1 , -1 ): heapify(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) for i in range(n - 1 , 0 , -1 ): lowerCamelCase_ , lowerCamelCase_ = array[0], array[i] heapify(lowerCamelCase__ , 0 , lowerCamelCase__ ) return array def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if (array[first_index] > array[middle_index]) != ( array[first_index] > array[last_index] ): return array[first_index] elif (array[middle_index] > array[first_index]) != ( array[middle_index] > array[last_index] ): return array[middle_index] else: return array[last_index] def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = low lowerCamelCase_ = high while True: while array[i] < pivot: i += 1 j -= 1 while pivot < array[j]: j -= 1 if i >= j: return i lowerCamelCase_ , lowerCamelCase_ = array[j], array[i] i += 1 def lowerCamelCase_ ( lowerCamelCase__ ): if len(lowerCamelCase__ ) == 0: return array lowerCamelCase_ = 2 * math.ceil(math.loga(len(lowerCamelCase__ ) ) ) lowerCamelCase_ = 1_6 return intro_sort(lowerCamelCase__ , 0 , len(lowerCamelCase__ ) , lowerCamelCase__ , lowerCamelCase__ ) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): while end - start > size_threshold: if max_depth == 0: return heap_sort(lowerCamelCase__ ) max_depth -= 1 lowerCamelCase_ = median_of_a(lowerCamelCase__ , lowerCamelCase__ , start + ((end - start) // 2) + 1 , end - 1 ) lowerCamelCase_ = partition(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) intro_sort(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ = p return insertion_sort(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) if __name__ == "__main__": import doctest doctest.testmod() __A =input('''Enter numbers separated by a comma : ''').strip() __A =[float(item) for item in user_input.split(''',''')] print(sort(unsorted))
19
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging __A =logging.get_logger(__name__) # pylint: disable=invalid-name class _SCREAMING_SNAKE_CASE ( snake_case_ ): def __init__( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , ) -> List[Any]: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: lowerCamelCase_ = ( f'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' f' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowercase , standard_warn=lowercase ) lowerCamelCase_ = dict(scheduler.config ) lowerCamelCase_ = 1 lowerCamelCase_ = FrozenDict(lowercase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: lowerCamelCase_ = ( f'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowercase , standard_warn=lowercase ) lowerCamelCase_ = dict(scheduler.config ) lowerCamelCase_ = True lowerCamelCase_ = FrozenDict(lowercase ) if safety_checker is None: logger.warning( f'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowercase , segmentation_processor=lowercase , vae=lowercase , text_encoder=lowercase , tokenizer=lowercase , unet=lowercase , scheduler=lowercase , safety_checker=lowercase , feature_extractor=lowercase , ) def SCREAMING_SNAKE_CASE_( self , lowercase = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory lowerCamelCase_ = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: self.enable_attention_slicing(lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) lowerCamelCase_ = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowercase , lowercase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self , lowercase , lowercase , lowercase , lowercase = 512 , lowercase = 512 , lowercase = 50 , lowercase = 7.5 , lowercase = None , lowercase = 1 , lowercase = 0.0 , lowercase = None , lowercase = None , lowercase = "pil" , lowercase = True , lowercase = None , lowercase = 1 , **lowercase , ) -> int: lowerCamelCase_ = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) lowerCamelCase_ = self.segmentation_model(**lowercase ) lowerCamelCase_ = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() lowerCamelCase_ = self.numpy_to_pil(lowercase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask lowerCamelCase_ = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowercase , image=lowercase , mask_image=lowercase , height=lowercase , width=lowercase , num_inference_steps=lowercase , guidance_scale=lowercase , negative_prompt=lowercase , num_images_per_prompt=lowercase , eta=lowercase , generator=lowercase , latents=lowercase , output_type=lowercase , return_dict=lowercase , callback=lowercase , callback_steps=lowercase , )
19
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a_ : List[str] = { 'configuration_graphormer': ['GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GraphormerConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : List[Any] = [ 'GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'GraphormerForGraphClassification', 'GraphormerModel', 'GraphormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys a_ : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
327
from math import isqrt def lowerCamelCase__ (_UpperCAmelCase): SCREAMING_SNAKE_CASE = [True] * max_number for i in range(2 , isqrt(max_number - 1) + 1): if is_prime[i]: for j in range(i**2 , _UpperCAmelCase , _UpperCAmelCase): SCREAMING_SNAKE_CASE = False return [i for i in range(2 , _UpperCAmelCase) if is_prime[i]] def lowerCamelCase__ (_UpperCAmelCase = 10**8): SCREAMING_SNAKE_CASE = calculate_prime_numbers(max_number // 2) SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = len(_UpperCAmelCase) - 1 while left <= right: while prime_numbers[left] * prime_numbers[right] >= max_number: right -= 1 semiprimes_count += right - left + 1 left += 1 return semiprimes_count if __name__ == "__main__": print(f"""{solution() = }""")
327
1
"""simple docstring""" import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel A: List[Any] = "0.12" # assumed parallelism: 8 @require_flax @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @classmethod def SCREAMING_SNAKE_CASE ( cls ) -> Any: '''simple docstring''' UpperCAmelCase : str = TOKEN HfFolder.save_token(_SCREAMING_SNAKE_CASE ) @classmethod def SCREAMING_SNAKE_CASE ( cls ) -> Dict: '''simple docstring''' try: delete_repo(token=cls._token , repo_id="""test-model-flax""" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="""valid_org/test-model-flax-org""" ) except HTTPError: pass def SCREAMING_SNAKE_CASE ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : Optional[Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Dict = FlaxBertModel(_SCREAMING_SNAKE_CASE ) model.push_to_hub("""test-model-flax""" , use_auth_token=self._token ) UpperCAmelCase : List[Any] = FlaxBertModel.from_pretrained(F"{USER}/test-model-flax" ) UpperCAmelCase : Union[str, Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Any = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Union[str, Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1E-3 , msg=F"{key} not identical" ) # Reset repo delete_repo(token=self._token , repo_id="""test-model-flax""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(_SCREAMING_SNAKE_CASE , repo_id="""test-model-flax""" , push_to_hub=_SCREAMING_SNAKE_CASE , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(F"{USER}/test-model-flax" ) UpperCAmelCase : int = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Dict = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Tuple = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1E-3 , msg=F"{key} not identical" ) def SCREAMING_SNAKE_CASE ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Optional[Any] = FlaxBertModel(_SCREAMING_SNAKE_CASE ) model.push_to_hub("""valid_org/test-model-flax-org""" , use_auth_token=self._token ) UpperCAmelCase : List[str] = FlaxBertModel.from_pretrained("""valid_org/test-model-flax-org""" ) UpperCAmelCase : str = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : List[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1E-3 , msg=F"{key} not identical" ) # Reset repo delete_repo(token=self._token , repo_id="""valid_org/test-model-flax-org""" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( _SCREAMING_SNAKE_CASE , repo_id="""valid_org/test-model-flax-org""" , push_to_hub=_SCREAMING_SNAKE_CASE , use_auth_token=self._token ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained("""valid_org/test-model-flax-org""" ) UpperCAmelCase : Dict = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Dict = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(_SCREAMING_SNAKE_CASE , 1E-3 , msg=F"{key} not identical" ) def _snake_case ( UpperCamelCase : Optional[Any] , UpperCamelCase : Dict ): UpperCAmelCase : Any = True UpperCAmelCase : Dict = flatten_dict(modela.params ) UpperCAmelCase : int = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: UpperCAmelCase : Tuple = False return models_are_equal @require_flax class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : List[str] = BertConfig.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) UpperCAmelCase : int = FlaxBertModel(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : List[Any] = """bert""" with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) with self.assertRaises(_SCREAMING_SNAKE_CASE ): UpperCAmelCase : Optional[int] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE , subfolder=_SCREAMING_SNAKE_CASE ) self.assertTrue(check_models_equal(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Dict = BertConfig.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) UpperCAmelCase : Tuple = FlaxBertModel(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Union[str, Any] = """bert""" with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) , max_shard_size="""10KB""" ) with self.assertRaises(_SCREAMING_SNAKE_CASE ): UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Optional[int] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE , subfolder=_SCREAMING_SNAKE_CASE ) self.assertTrue(check_models_equal(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) def SCREAMING_SNAKE_CASE ( self ) -> List[str]: '''simple docstring''' UpperCAmelCase : int = """bert""" UpperCAmelCase : Union[str, Any] = """hf-internal-testing/tiny-random-bert-subfolder""" with self.assertRaises(_SCREAMING_SNAKE_CASE ): UpperCAmelCase : Optional[Any] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Optional[int] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE , subfolder=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' UpperCAmelCase : Tuple = """bert""" UpperCAmelCase : Dict = """hf-internal-testing/tiny-random-bert-sharded-subfolder""" with self.assertRaises(_SCREAMING_SNAKE_CASE ): UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE ) UpperCAmelCase : Optional[int] = FlaxBertModel.from_pretrained(_SCREAMING_SNAKE_CASE , subfolder=_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE )
109
"""simple docstring""" import argparse import torch from transformers import BertForMaskedLM if __name__ == "__main__": _a = argparse.ArgumentParser( description=( """Extraction some layers of the full BertForMaskedLM or RObertaForMaskedLM for Transfer Learned""" """ Distillation""" ) ) parser.add_argument("""--model_type""", default="""bert""", choices=["""bert"""]) parser.add_argument("""--model_name""", default="""bert-base-uncased""", type=str) parser.add_argument("""--dump_checkpoint""", default="""serialization_dir/tf_bert-base-uncased_0247911.pth""", type=str) parser.add_argument("""--vocab_transform""", action="""store_true""") _a = parser.parse_args() if args.model_type == "bert": _a = BertForMaskedLM.from_pretrained(args.model_name) _a = """bert""" else: raise ValueError("""args.model_type should be \"bert\".""") _a = model.state_dict() _a = {} for w in ["word_embeddings", "position_embeddings"]: _a = state_dict[F"""{prefix}.embeddings.{w}.weight"""] for w in ["weight", "bias"]: _a = state_dict[F"""{prefix}.embeddings.LayerNorm.{w}"""] _a = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: for w in ["weight", "bias"]: _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.attention.self.query.{w}""" ] _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.attention.self.key.{w}""" ] _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.attention.self.value.{w}""" ] _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.attention.output.dense.{w}""" ] _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.attention.output.LayerNorm.{w}""" ] _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.intermediate.dense.{w}""" ] _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.output.dense.{w}""" ] _a = state_dict[ F"""{prefix}.encoder.layer.{teacher_idx}.output.LayerNorm.{w}""" ] std_idx += 1 _a = state_dict["""cls.predictions.decoder.weight"""] _a = state_dict["""cls.predictions.bias"""] if args.vocab_transform: for w in ["weight", "bias"]: _a = state_dict[F"""cls.predictions.transform.dense.{w}"""] _a = state_dict[F"""cls.predictions.transform.LayerNorm.{w}"""] print(F"""N layers selected for distillation: {std_idx}""") print(F"""Number of params transferred for distillation: {len(compressed_sd.keys())}""") print(F"""Save transferred checkpoint to {args.dump_checkpoint}.""") torch.save(compressed_sd, args.dump_checkpoint)
194
0
"""simple docstring""" from __future__ import annotations def UpperCAmelCase__ (lowerCAmelCase_ ): '''simple docstring''' for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(lowerCAmelCase_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(lowerCAmelCase_ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
195
"""simple docstring""" import numpy as np import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModelWithProjection, PreTrainedModel from ...utils import logging a__ : Optional[Any] = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase): """simple docstring""" snake_case__ : int = CLIPConfig snake_case__ : str = ["CLIPEncoderLayer"] def __init__( self : Optional[int] , UpperCAmelCase__ : CLIPConfig ) -> Dict: super().__init__(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = CLIPVisionModelWithProjection(config.vision_config ) __SCREAMING_SNAKE_CASE = nn.Linear(config.vision_config.projection_dim , 1 ) __SCREAMING_SNAKE_CASE = nn.Linear(config.vision_config.projection_dim , 1 ) @torch.no_grad() def UpperCAmelCase_ ( self : str , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : int=0.5 , UpperCAmelCase__ : Optional[int]=0.5 ) -> Union[str, Any]: __SCREAMING_SNAKE_CASE = self.vision_model(UpperCAmelCase__ )[0] __SCREAMING_SNAKE_CASE = self.p_head(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = nsfw_detected.flatten() __SCREAMING_SNAKE_CASE = nsfw_detected > p_threshold __SCREAMING_SNAKE_CASE = nsfw_detected.tolist() if any(UpperCAmelCase__ ): logger.warning( "Potential NSFW content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed." ) for idx, nsfw_detected_ in enumerate(UpperCAmelCase__ ): if nsfw_detected_: __SCREAMING_SNAKE_CASE = np.zeros(images[idx].shape ) __SCREAMING_SNAKE_CASE = self.w_head(UpperCAmelCase__ ) __SCREAMING_SNAKE_CASE = watermark_detected.flatten() __SCREAMING_SNAKE_CASE = watermark_detected > w_threshold __SCREAMING_SNAKE_CASE = watermark_detected.tolist() if any(UpperCAmelCase__ ): logger.warning( "Potential watermarked content was detected in one or more images. A black image will be returned instead." " Try again with a different prompt and/or seed." ) for idx, watermark_detected_ in enumerate(UpperCAmelCase__ ): if watermark_detected_: __SCREAMING_SNAKE_CASE = np.zeros(images[idx].shape ) return images, nsfw_detected, watermark_detected
195
1
import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors __snake_case : Union[str, Any] =logging.getLogger(__name__) class lowerCamelCase__ ( lowerCamelCase__): '''simple docstring''' snake_case_ ="""sequence-classification""" def __init__(self ,__lowerCamelCase ) -> Union[str, Any]: """simple docstring""" if type(__lowercase ) == dict: lowerCAmelCase__ : Optional[Any] = Namespace(**__lowercase ) lowerCAmelCase__ : Any = glue_output_modes[hparams.task] lowerCAmelCase__ : Union[str, Any] = glue_tasks_num_labels[hparams.task] super().__init__(__lowercase ,__lowercase ,self.mode ) def lowerCAmelCase__ (self ,**__lowerCamelCase ) -> Any: """simple docstring""" return self.model(**__lowercase ) def lowerCAmelCase__ (self ,__lowerCamelCase ,__lowerCamelCase ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Optional[Any] = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowerCAmelCase__ : Dict = batch[2] if self.config.model_type in ['''bert''', '''xlnet''', '''albert'''] else None lowerCAmelCase__ : Optional[int] = self(**__lowercase ) lowerCAmelCase__ : Tuple = outputs[0] lowerCAmelCase__ : int = self.trainer.lr_schedulers[0]['''scheduler'''] lowerCAmelCase__ : Dict = {'''loss''': loss, '''rate''': lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def lowerCAmelCase__ (self ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Tuple = self.hparams lowerCAmelCase__ : Dict = processors[args.task]() lowerCAmelCase__ : Union[str, Any] = processor.get_labels() for mode in ["train", "dev"]: lowerCAmelCase__ : Dict = self._feature_file(__lowercase ) if os.path.exists(__lowercase ) and not args.overwrite_cache: logger.info('''Loading features from cached file %s''' ,__lowercase ) else: logger.info('''Creating features from dataset file at %s''' ,args.data_dir ) lowerCAmelCase__ : Optional[Any] = ( processor.get_dev_examples(args.data_dir ) if mode == '''dev''' else processor.get_train_examples(args.data_dir ) ) lowerCAmelCase__ : Optional[Any] = convert_examples_to_features( __lowercase ,self.tokenizer ,max_length=args.max_seq_length ,label_list=self.labels ,output_mode=args.glue_output_mode ,) logger.info('''Saving features into cached file %s''' ,__lowercase ) torch.save(__lowercase ,__lowercase ) def lowerCAmelCase__ (self ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase = False ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : Optional[int] = '''dev''' if mode == '''test''' else mode lowerCAmelCase__ : Any = self._feature_file(__lowercase ) logger.info('''Loading features from cached file %s''' ,__lowercase ) lowerCAmelCase__ : Dict = torch.load(__lowercase ) lowerCAmelCase__ : Dict = torch.tensor([f.input_ids for f in features] ,dtype=torch.long ) lowerCAmelCase__ : Optional[int] = torch.tensor([f.attention_mask for f in features] ,dtype=torch.long ) lowerCAmelCase__ : Dict = torch.tensor([f.token_type_ids for f in features] ,dtype=torch.long ) if self.hparams.glue_output_mode == "classification": lowerCAmelCase__ : Optional[Any] = torch.tensor([f.label for f in features] ,dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": lowerCAmelCase__ : str = torch.tensor([f.label for f in features] ,dtype=torch.float ) return DataLoader( TensorDataset(__lowercase ,__lowercase ,__lowercase ,__lowercase ) ,batch_size=__lowercase ,shuffle=__lowercase ,) def lowerCAmelCase__ (self ,__lowerCamelCase ,__lowerCamelCase ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : List[Any] = {'''input_ids''': batch[0], '''attention_mask''': batch[1], '''labels''': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: lowerCAmelCase__ : List[Any] = batch[2] if self.config.model_type in ['''bert''', '''xlnet''', '''albert'''] else None lowerCAmelCase__ : Optional[int] = self(**__lowercase ) lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = outputs[:2] lowerCAmelCase__ : List[str] = logits.detach().cpu().numpy() lowerCAmelCase__ : int = inputs['''labels'''].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def lowerCAmelCase__ (self ,__lowerCamelCase ) -> Dict: """simple docstring""" lowerCAmelCase__ : List[Any] = torch.stack([x['''val_loss'''] for x in outputs] ).mean().detach().cpu().item() lowerCAmelCase__ : List[Any] = np.concatenate([x['''pred'''] for x in outputs] ,axis=0 ) if self.hparams.glue_output_mode == "classification": lowerCAmelCase__ : List[str] = np.argmax(__lowercase ,axis=1 ) elif self.hparams.glue_output_mode == "regression": lowerCAmelCase__ : List[Any] = np.squeeze(__lowercase ) lowerCAmelCase__ : Union[str, Any] = np.concatenate([x['''target'''] for x in outputs] ,axis=0 ) lowerCAmelCase__ : int = [[] for _ in range(out_label_ids.shape[0] )] lowerCAmelCase__ : Optional[Any] = [[] for _ in range(out_label_ids.shape[0] )] lowerCAmelCase__ : Tuple = {**{'''val_loss''': val_loss_mean}, **compute_metrics(self.hparams.task ,__lowercase ,__lowercase )} lowerCAmelCase__ : Union[str, Any] = dict(results.items() ) lowerCAmelCase__ : List[Any] = results return ret, preds_list, out_label_list def lowerCAmelCase__ (self ,__lowerCamelCase ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : List[str] = self._eval_end(__lowercase ) lowerCAmelCase__ : List[str] = ret['''log'''] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def lowerCAmelCase__ (self ,__lowerCamelCase ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ : str = self._eval_end(__lowercase ) lowerCAmelCase__ : int = ret['''log'''] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def lowerCAmelCase__ (__lowerCamelCase ,__lowerCamelCase ) -> Optional[Any]: """simple docstring""" BaseTransformer.add_model_specific_args(__lowercase ,__lowercase ) parser.add_argument( '''--max_seq_length''' ,default=1_28 ,type=__lowercase ,help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) ,) parser.add_argument( '''--task''' ,default='''''' ,type=__lowercase ,required=__lowercase ,help='''The GLUE task to run''' ,) parser.add_argument( '''--gpus''' ,default=0 ,type=__lowercase ,help='''The number of GPUs allocated for this, it is by default 0 meaning none''' ,) parser.add_argument( '''--overwrite_cache''' ,action='''store_true''' ,help='''Overwrite the cached training and evaluation sets''' ) return parser def lowerCAmelCase__ ( ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = argparse.ArgumentParser() add_generic_args(lowercase__ ,os.getcwd()) lowerCAmelCase__ : Dict = GLUETransformer.add_model_specific_args(lowercase__ ,os.getcwd()) lowerCAmelCase__ : Optional[int] = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: lowerCAmelCase__ : Dict = os.path.join( '''./results''' ,f"""{args.task}_{time.strftime('%Y%m%d_%H%M%S')}""" ,) os.makedirs(args.output_dir) lowerCAmelCase__ : Dict = GLUETransformer(lowercase__) lowerCAmelCase__ : Union[str, Any] = generic_train(lowercase__ ,lowercase__) # Optionally, predict on dev set and write to output_dir if args.do_predict: lowerCAmelCase__ : str = sorted(glob.glob(os.path.join(args.output_dir ,'''checkpoint-epoch=*.ckpt''') ,recursive=lowercase__)) lowerCAmelCase__ : Optional[Any] = model.load_from_checkpoint(checkpoints[-1]) return trainer.test(lowercase__) if __name__ == "__main__": main()
129
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("TEST_SAGEMAKER" , "False" ) ) is not True , reason="Skipping test because should only be run when releasing minor transformers version" , ) @pytest.mark.usefixtures("sm_env" ) @parameterized_class( [ { "framework": "pytorch", "script": "run_glue.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_5_0, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "pytorch", "script": "run_ddp.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_0_0, "eval_accuracy": 0.7, "eval_loss": 0.6}, }, { "framework": "tensorflow", "script": "run_tf_dist.py", "model_name_or_path": "distilbert-base-cased", "instance_type": "ml.p3.16xlarge", "results": {"train_runtime": 6_0_0, "eval_accuracy": 0.6, "eval_loss": 0.7}, }, ] ) class lowerCAmelCase ( unittest.TestCase ): def snake_case ( self : int ): """simple docstring""" if self.framework == "pytorch": subprocess.run( f'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding='utf-8' , check=__lowercase , ) assert hasattr(self , 'env' ) def snake_case ( self : Tuple , __lowercase : List[str] ): """simple docstring""" __lowercase =f'''{self.env.base_job_name}-{instance_count}-{"ddp" if "ddp" in self.script else "smd"}''' # distributed data settings __lowercase ={'smdistributed': {'dataparallel': {'enabled': True}}} if self.script != 'run_ddp.py' else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=__lowercase , instance_count=__lowercase , instance_type=self.instance_type , debugger_hook_config=__lowercase , hyperparameters={**self.env.distributed_hyperparameters, 'model_name_or_path': self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=__lowercase , py_version='py36' , ) def snake_case ( self : int , __lowercase : List[str] ): """simple docstring""" TrainingJobAnalytics(__lowercase ).export_csv(f'''{self.env.test_path}/{job_name}_metrics.csv''' ) @parameterized.expand([(2,)] ) def snake_case ( self : Tuple , __lowercase : List[Any] ): """simple docstring""" __lowercase =self.create_estimator(__lowercase ) # run training estimator.fit() # result dataframe __lowercase =TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __lowercase =list(result_metrics_df[result_metrics_df.metric_name == 'eval_accuracy']['value'] ) __lowercase =list(result_metrics_df[result_metrics_df.metric_name == 'eval_loss']['value'] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __lowercase =( Session().describe_training_job(estimator.latest_training_job.name ).get('TrainingTimeInSeconds' , 999999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['eval_accuracy'] for t in eval_accuracy ) assert all(t <= self.results['eval_loss'] for t in eval_loss ) # dump tests result into json file to share in PR with open(f'''{estimator.latest_training_job.name}.json''' , 'w' ) as outfile: json.dump({'train_time': train_runtime, 'eval_accuracy': eval_accuracy, 'eval_loss': eval_loss} , __lowercase )
141
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__ = { """configuration_rembert""": ["""REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RemBertConfig""", """RemBertOnnxConfig"""] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ["""RemBertTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ["""RemBertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ """REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """RemBertForCausalLM""", """RemBertForMaskedLM""", """RemBertForMultipleChoice""", """RemBertForQuestionAnswering""", """RemBertForSequenceClassification""", """RemBertForTokenClassification""", """RemBertLayer""", """RemBertModel""", """RemBertPreTrainedModel""", """load_tf_weights_in_rembert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ """TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRemBertForCausalLM""", """TFRemBertForMaskedLM""", """TFRemBertForMultipleChoice""", """TFRemBertForQuestionAnswering""", """TFRemBertForSequenceClassification""", """TFRemBertForTokenClassification""", """TFRemBertLayer""", """TFRemBertModel""", """TFRemBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
102
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCamelCase__ = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class a__ : _a : str = field( default=snake_case__ , metadata={"""help""": """Model type selected in the list: """ + """, """.join(snake_case__ )} ) _a : str = field( default=snake_case__ , metadata={"""help""": """The input data dir. Should contain the .json files for the SQuAD task."""} ) _a : int = field( default=1_2_8 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) _a : int = field( default=1_2_8 , metadata={"""help""": """When splitting up a long document into chunks, how much stride to take between chunks."""} , ) _a : int = field( default=6_4 , metadata={ """help""": ( """The maximum number of tokens for the question. Questions longer than this will """ """be truncated to this length.""" ) } , ) _a : int = field( default=3_0 , metadata={ """help""": ( """The maximum length of an answer that can be generated. This is needed because the start """ """and end predictions are not conditioned on one another.""" ) } , ) _a : bool = field( default=snake_case__ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) _a : bool = field( default=snake_case__ , metadata={"""help""": """If true, the SQuAD examples contain some that do not have an answer."""} ) _a : float = field( default=0.0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _a : int = field( default=2_0 , metadata={"""help""": """If null_score - best_non_null is greater than the threshold predict null."""} ) _a : int = field( default=0 , metadata={ """help""": ( """language id of input for language-specific xlm models (see""" """ tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)""" ) } , ) _a : int = field(default=1 , metadata={"""help""": """multiple threads for converting example to features"""} ) class a__ ( snake_case__ ): _a : Any = """train""" _a : Union[str, Any] = """dev""" class a__ ( snake_case__ ): _a : SquadDataTrainingArguments _a : List[SquadFeatures] _a : Split _a : bool def __init__( self , _A , _A , _A = None , _A = Split.train , _A = False , _A = None , _A = "pt" , ): """simple docstring""" __lowerCAmelCase = args __lowerCAmelCase = is_language_sensitive __lowerCAmelCase = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(_A , _A ): try: __lowerCAmelCase = Split[mode] except KeyError: raise KeyError("mode is not a valid split name" ) __lowerCAmelCase = mode # Load data features from cache or dataset file __lowerCAmelCase = "v2" if args.version_2_with_negative else "v1" __lowerCAmelCase = os.path.join( cache_dir if cache_dir is not None else args.data_dir , f"""cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}""" , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __lowerCAmelCase = cached_features_file + ".lock" with FileLock(_A ): if os.path.exists(_A ) and not args.overwrite_cache: __lowerCAmelCase = time.time() __lowerCAmelCase = torch.load(_A ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. __lowerCAmelCase = self.old_features["features"] __lowerCAmelCase = self.old_features.get("dataset" , _A ) __lowerCAmelCase = self.old_features.get("examples" , _A ) logger.info( f"""Loading features from cached file {cached_features_file} [took %.3f s]""" , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( f"""Deleting cached file {cached_features_file} will allow dataset and examples to be cached in""" " future run" ) else: if mode == Split.dev: __lowerCAmelCase = self.processor.get_dev_examples(args.data_dir ) else: __lowerCAmelCase = self.processor.get_train_examples(args.data_dir ) __lowerCAmelCase , __lowerCAmelCase = squad_convert_examples_to_features( examples=self.examples , tokenizer=_A , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=_A , ) __lowerCAmelCase = time.time() torch.save( {"features": self.features, "dataset": self.dataset, "examples": self.examples} , _A , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( f"""Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]""" ) def __len__( self ): """simple docstring""" return len(self.features ) def __getitem__( self , _A ): """simple docstring""" __lowerCAmelCase = self.features[i] __lowerCAmelCase = torch.tensor(feature.input_ids , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.attention_mask , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.token_type_ids , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.cls_index , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.p_mask , dtype=torch.float ) __lowerCAmelCase = torch.tensor(feature.is_impossible , dtype=torch.float ) __lowerCAmelCase = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"cls_index": cls_index, "p_mask": p_mask} ) if self.args.version_2_with_negative: inputs.update({"is_impossible": is_impossible} ) if self.is_language_sensitive: inputs.update({"langs": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: __lowerCAmelCase = torch.tensor(feature.start_position , dtype=torch.long ) __lowerCAmelCase = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"start_positions": start_positions, "end_positions": end_positions} ) return inputs
102
1
"""simple docstring""" def A_ ( _lowercase = 100 ): '''simple docstring''' snake_case_ :Dict = set() snake_case_ :Tuple = 0 snake_case_ :Optional[int] = n + 1 # maximum limit for a in range(2, _lowercase ): for b in range(2, _lowercase ): snake_case_ :Optional[Any] = a**b # calculates the current power collect_powers.add(_lowercase ) # adds the result to the set return len(_lowercase ) if __name__ == "__main__": print("Number of terms ", solution(int(str(input()).strip())))
66
"""simple docstring""" import unittest import numpy as np from transformers import BertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.bert.modeling_flax_bert import ( FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, ) class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self: List[Any] , snake_case: List[str] , snake_case: Optional[Any]=13 , snake_case: List[str]=7 , snake_case: Dict=True , snake_case: List[str]=True , snake_case: Optional[int]=True , snake_case: Any=True , snake_case: Optional[Any]=99 , snake_case: Tuple=32 , snake_case: Tuple=5 , snake_case: Dict=4 , snake_case: Optional[Any]=37 , snake_case: Union[str, Any]="gelu" , snake_case: Tuple=0.1 , snake_case: List[Any]=0.1 , snake_case: List[str]=512 , snake_case: Optional[int]=16 , snake_case: int=2 , snake_case: List[Any]=0.0_2 , snake_case: Union[str, Any]=4 , ) -> List[str]: snake_case_ :Dict = parent snake_case_ :Any = batch_size snake_case_ :Any = seq_length snake_case_ :List[str] = is_training snake_case_ :Optional[Any] = use_attention_mask snake_case_ :Dict = use_token_type_ids snake_case_ :Union[str, Any] = use_labels snake_case_ :str = vocab_size snake_case_ :int = hidden_size snake_case_ :List[str] = num_hidden_layers snake_case_ :Dict = num_attention_heads snake_case_ :Any = intermediate_size snake_case_ :Tuple = hidden_act snake_case_ :int = hidden_dropout_prob snake_case_ :Optional[Any] = attention_probs_dropout_prob snake_case_ :Any = max_position_embeddings snake_case_ :Union[str, Any] = type_vocab_size snake_case_ :Optional[int] = type_sequence_label_size snake_case_ :Union[str, Any] = initializer_range snake_case_ :Tuple = num_choices def lowerCAmelCase_ ( self: Tuple ) -> str: snake_case_ :Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ :Union[str, Any] = None if self.use_attention_mask: snake_case_ :str = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ :Any = None if self.use_token_type_ids: snake_case_ :List[str] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ :int = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowerCAmelCase_ ( self: Optional[int] ) -> int: snake_case_ :str = self.prepare_config_and_inputs() snake_case_, snake_case_, snake_case_, snake_case_ :Optional[int] = config_and_inputs snake_case_ :Union[str, Any] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def lowerCAmelCase_ ( self: Optional[Any] ) -> Any: snake_case_ :int = self.prepare_config_and_inputs() snake_case_, snake_case_, snake_case_, snake_case_ :Dict = config_and_inputs snake_case_ :Union[str, Any] = True snake_case_ :Optional[int] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ :Tuple = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, attention_mask, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowerCamelCase ( _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' _A : List[str] = True _A : Dict = ( ( FlaxBertModel, FlaxBertForPreTraining, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForQuestionAnswering, FlaxBertForNextSentencePrediction, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertForQuestionAnswering, ) if is_flax_available() else () ) def lowerCAmelCase_ ( self: int ) -> List[str]: snake_case_ :Any = FlaxBertModelTester(self ) @slow def lowerCAmelCase_ ( self: List[str] ) -> Dict: # Only check this for base model, not necessary for all model classes. # This will also help speed-up tests. snake_case_ :Dict = FlaxBertModel.from_pretrained("""bert-base-cased""" ) snake_case_ :Dict = model(np.ones((1, 1) ) ) self.assertIsNotNone(snake_case )
66
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> int: '''simple docstring''' if n == 1 or not isinstance(__lowerCAmelCase , __lowerCAmelCase ): return 0 elif n == 2: return 1 else: lowercase_ = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> int: '''simple docstring''' lowercase_ = 0 lowercase_ = 2 while digits < n: index += 1 lowercase_ = len(str(fibonacci(__lowerCAmelCase ) ) ) return index def _SCREAMING_SNAKE_CASE (__lowerCAmelCase = 10_00 ) -> int: '''simple docstring''' return fibonacci_digits_index(__lowerCAmelCase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
366
"""simple docstring""" from __future__ import annotations def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> list[str]: '''simple docstring''' if nth_term == "": return [""] lowercase_ = int(__lowerCAmelCase ) lowercase_ = int(__lowerCAmelCase ) lowercase_ = [] for temp in range(int(__lowerCAmelCase ) ): series.append(F'''1 / {pow(temp + 1 , int(__lowerCAmelCase ) )}''' if series else """1""" ) return series if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase : List[str] = int(input("Enter the last number (nth term) of the P-Series")) UpperCAmelCase : Tuple = int(input("Enter the power for P-Series")) print("Formula of P-Series => 1+1/2^p+1/3^p ..... 1/n^p") print(p_series(nth_term, power))
313
0
from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a__ : """simple docstring""" def __init__( self : Optional[Any] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[Any]=3 , UpperCAmelCase__ : int=3_2 , UpperCAmelCase__ : Any=3 , UpperCAmelCase__ : List[str]=1_0 , UpperCAmelCase__ : Tuple=[1_0, 2_0, 3_0, 4_0] , UpperCAmelCase__ : Optional[int]=[1, 1, 2, 1] , UpperCAmelCase__ : List[str]=True , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Dict="relu" , UpperCAmelCase__ : Tuple=3 , UpperCAmelCase__ : Dict=None , ) ->Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : int = parent SCREAMING_SNAKE_CASE : Dict = batch_size SCREAMING_SNAKE_CASE : Union[str, Any] = image_size SCREAMING_SNAKE_CASE : Any = num_channels SCREAMING_SNAKE_CASE : Dict = embeddings_size SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_sizes SCREAMING_SNAKE_CASE : Any = depths SCREAMING_SNAKE_CASE : List[Any] = is_training SCREAMING_SNAKE_CASE : List[Any] = use_labels SCREAMING_SNAKE_CASE : Any = hidden_act SCREAMING_SNAKE_CASE : int = num_labels SCREAMING_SNAKE_CASE : Optional[Any] = scope SCREAMING_SNAKE_CASE : Dict = len(lowerCAmelCase__ ) def _lowercase ( self : List[str] ) ->List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE : Tuple = None if self.use_labels: SCREAMING_SNAKE_CASE : Tuple = ids_tensor([self.batch_size] , self.num_labels ) SCREAMING_SNAKE_CASE : int = self.get_config() return config, pixel_values, labels def _lowercase ( self : Dict ) ->Optional[int]: """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : int ) ->str: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = TFRegNetModel(config=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[int] = model(lowerCAmelCase__ , training=lowerCAmelCase__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def _lowercase ( self : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Tuple ) ->str: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = self.num_labels SCREAMING_SNAKE_CASE : Any = TFRegNetForImageClassification(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Any = model(lowerCAmelCase__ , labels=lowerCAmelCase__ , training=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : List[Any] ) ->Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE : Tuple = config_and_inputs SCREAMING_SNAKE_CASE : Optional[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class a__ ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" UpperCAmelCase__ : Any =(TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () UpperCAmelCase__ : Optional[int] =( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) UpperCAmelCase__ : List[str] =False UpperCAmelCase__ : List[str] =False UpperCAmelCase__ : Optional[Any] =False UpperCAmelCase__ : Tuple =False UpperCAmelCase__ : int =False def _lowercase ( self : Any ) ->Dict: """simple docstring""" SCREAMING_SNAKE_CASE : Any = TFRegNetModelTester(self ) SCREAMING_SNAKE_CASE : int = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ ) def _lowercase ( self : Any ) ->Union[str, Any]: """simple docstring""" return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def _lowercase ( self : Any ) ->Tuple: """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def _lowercase ( self : Dict ) ->Any: """simple docstring""" super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def _lowercase ( self : Union[str, Any] ) ->List[Any]: """simple docstring""" pass def _lowercase ( self : Tuple ) ->Tuple: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : Tuple = model_class(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE : List[str] = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def _lowercase ( self : Optional[Any] ) ->List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def _lowercase ( self : int ) ->str: """simple docstring""" def check_hidden_states_output(UpperCAmelCase__ : str , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : int ): SCREAMING_SNAKE_CASE : Optional[int] = model_class(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Tuple = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) , training=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states SCREAMING_SNAKE_CASE : Optional[Any] = self.model_tester.num_stages self.assertEqual(len(lowerCAmelCase__ ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) SCREAMING_SNAKE_CASE : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE : List[Any] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: SCREAMING_SNAKE_CASE : Optional[int] = layer_type SCREAMING_SNAKE_CASE : Optional[Any] = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE : Any = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self : List[Any] ) ->Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(UpperCAmelCase__ : int , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Any={} ): SCREAMING_SNAKE_CASE : Any = model(lowerCAmelCase__ , return_dict=lowerCAmelCase__ , **lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = model(lowerCAmelCase__ , return_dict=lowerCAmelCase__ , **lowerCAmelCase__ ).to_tuple() def recursive_check(UpperCAmelCase__ : Dict , UpperCAmelCase__ : str ): if isinstance(lowerCAmelCase__ , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(lowerCAmelCase__ , lowerCAmelCase__ ): recursive_check(lowerCAmelCase__ , lowerCAmelCase__ ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(lowerCAmelCase__ , lowerCAmelCase__ ) ) , msg=( """Tuple and dict output are not equal. Difference:""" f" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}" ) , ) recursive_check(lowerCAmelCase__ , lowerCAmelCase__ ) for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE : int = model_class(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Optional[Any] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : List[str] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) check_equivalence(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Any = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) check_equivalence(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : str = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : Union[str, Any] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) check_equivalence(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , {"""output_hidden_states""": True} ) SCREAMING_SNAKE_CASE : List[str] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : List[Any] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) check_equivalence(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , {"""output_hidden_states""": True} ) def _lowercase ( self : Tuple ) ->str: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) @slow def _lowercase ( self : Dict ) ->Optional[int]: """simple docstring""" for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE : Union[str, Any] = TFRegNetModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def __lowercase ( ) -> int: SCREAMING_SNAKE_CASE : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class a__ ( unittest.TestCase ): """simple docstring""" @cached_property def _lowercase ( self : Optional[Any] ) ->Optional[int]: """simple docstring""" return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def _lowercase ( self : int ) ->Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) SCREAMING_SNAKE_CASE : Dict = self.default_image_processor SCREAMING_SNAKE_CASE : str = prepare_img() SCREAMING_SNAKE_CASE : Dict = image_processor(images=lowerCAmelCase__ , return_tensors="""tf""" ) # forward pass SCREAMING_SNAKE_CASE : Any = model(**lowerCAmelCase__ , training=lowerCAmelCase__ ) # verify the logits SCREAMING_SNAKE_CASE : Any = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE : int = tf.constant([-0.41_80, -1.50_51, -3.48_36] ) tf.debugging.assert_near(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1e-4 )
245
import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class snake_case__ ( _lowerCAmelCase , unittest.TestCase ): lowercase__ : Optional[Any] = MgpstrTokenizer lowercase__ : int = False lowercase__ : Any = {} lowercase__ : Optional[int] = False def __magic_name__ ( self ) -> Optional[Any]: super().setUp() # fmt: off __magic_name__ : List[str] = ["""[GO]""", """[s]""", """0""", """1""", """2""", """3""", """4""", """5""", """6""", """7""", """8""", """9""", """a""", """b""", """c""", """d""", """e""", """f""", """g""", """h""", """i""", """j""", """k""", """l""", """m""", """n""", """o""", """p""", """q""", """r""", """s""", """t""", """u""", """v""", """w""", """x""", """y""", """z"""] # fmt: on __magic_name__ : List[Any] = dict(zip(lowerCAmelCase__ , range(len(lowerCAmelCase__ ) ) ) ) __magic_name__ : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + """\n""" ) def __magic_name__ ( self , **lowerCAmelCase__ ) -> Optional[int]: return MgpstrTokenizer.from_pretrained(self.tmpdirname , **lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Optional[int]: __magic_name__ : List[str] = """tester""" __magic_name__ : int = """tester""" return input_text, output_text @unittest.skip("""MGP-STR always lower cases letters.""" ) def __magic_name__ ( self ) -> str: pass def __magic_name__ ( self ) -> List[str]: __magic_name__ : List[Any] = self.get_tokenizers(do_lower_case=lowerCAmelCase__ ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): __magic_name__ : Dict = """[SPECIAL_TOKEN]""" tokenizer.add_special_tokens({"""cls_token""": special_token} ) __magic_name__ : List[str] = tokenizer.encode([special_token] , add_special_tokens=lowerCAmelCase__ ) self.assertEqual(len(lowerCAmelCase__ ) , 1 ) __magic_name__ : Tuple = tokenizer.decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) self.assertTrue(special_token not in decoded ) def __magic_name__ ( self ) -> Union[str, Any]: __magic_name__ : int = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): __magic_name__ ,__magic_name__ : Optional[Any] = self.get_input_output_texts(lowerCAmelCase__ ) __magic_name__ : List[Any] = tokenizer.tokenize(lowerCAmelCase__ ) __magic_name__ : Any = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) __magic_name__ : Union[str, Any] = tokenizer.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) __magic_name__ : List[Any] = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertNotEqual(len(lowerCAmelCase__ ) , 0 ) __magic_name__ : Optional[int] = tokenizer.decode(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertEqual(text_a.replace(""" """ , """""" ) , lowerCAmelCase__ ) @unittest.skip("""MGP-STR tokenizer only handles one sequence.""" ) def __magic_name__ ( self ) -> Tuple: pass @unittest.skip("""inputs cannot be pretokenized in MgpstrTokenizer""" ) def __magic_name__ ( self ) -> Optional[Any]: pass
342
0
"""simple docstring""" import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BertTokenizer, BlipImageProcessor, BlipProcessor, PreTrainedTokenizerFast @require_vision class _UpperCAmelCase( unittest.TestCase ): def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' _UpperCamelCase = tempfile.mkdtemp() _UpperCamelCase = BlipImageProcessor() _UpperCamelCase = BertTokenizer.from_pretrained('''hf-internal-testing/tiny-random-BertModel''') _UpperCamelCase = BlipProcessor(__a , __a) processor.save_pretrained(self.tmpdirname) def UpperCAmelCase ( self , **__a) -> Tuple: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **__a).tokenizer def UpperCAmelCase ( self , **__a) -> Any: '''simple docstring''' return AutoProcessor.from_pretrained(self.tmpdirname , **__a).image_processor def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' shutil.rmtree(self.tmpdirname) def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = [np.random.randint(2_55 , size=(3, 30, 4_00) , dtype=np.uinta)] _UpperCamelCase = [Image.fromarray(np.moveaxis(__a , 0 , -1)) for x in image_inputs] return image_inputs def UpperCAmelCase ( self) -> Optional[int]: '''simple docstring''' _UpperCamelCase = BlipProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor()) processor.save_pretrained(self.tmpdirname) _UpperCamelCase = self.get_tokenizer(bos_token='''(BOS)''' , eos_token='''(EOS)''') _UpperCamelCase = self.get_image_processor(do_normalize=__a , padding_value=1.0) _UpperCamelCase = BlipProcessor.from_pretrained( self.tmpdirname , bos_token='''(BOS)''' , eos_token='''(EOS)''' , do_normalize=__a , padding_value=1.0) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab()) self.assertIsInstance(processor.tokenizer , __a) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string()) self.assertIsInstance(processor.image_processor , __a) def UpperCAmelCase ( self) -> List[str]: '''simple docstring''' _UpperCamelCase = self.get_image_processor() _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = BlipProcessor(tokenizer=__a , image_processor=__a) _UpperCamelCase = self.prepare_image_inputs() _UpperCamelCase = image_processor(__a , return_tensors='''np''') _UpperCamelCase = processor(images=__a , return_tensors='''np''') for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2) def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' _UpperCamelCase = self.get_image_processor() _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = BlipProcessor(tokenizer=__a , image_processor=__a) _UpperCamelCase = '''lower newer''' _UpperCamelCase = processor(text=__a) _UpperCamelCase = tokenizer(__a , return_token_type_ids=__a) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key]) def UpperCAmelCase ( self) -> Tuple: '''simple docstring''' _UpperCamelCase = self.get_image_processor() _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = BlipProcessor(tokenizer=__a , image_processor=__a) _UpperCamelCase = '''lower newer''' _UpperCamelCase = self.prepare_image_inputs() _UpperCamelCase = processor(text=__a , images=__a) self.assertListEqual(list(inputs.keys()) , ['''pixel_values''', '''input_ids''', '''attention_mask''']) # test if it raises when no input is passed with pytest.raises(__a): processor() def UpperCAmelCase ( self) -> str: '''simple docstring''' _UpperCamelCase = self.get_image_processor() _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = BlipProcessor(tokenizer=__a , image_processor=__a) _UpperCamelCase = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] _UpperCamelCase = processor.batch_decode(__a) _UpperCamelCase = tokenizer.batch_decode(__a) self.assertListEqual(__a , __a) def UpperCAmelCase ( self) -> Any: '''simple docstring''' _UpperCamelCase = self.get_image_processor() _UpperCamelCase = self.get_tokenizer() _UpperCamelCase = BlipProcessor(tokenizer=__a , image_processor=__a) _UpperCamelCase = '''lower newer''' _UpperCamelCase = self.prepare_image_inputs() _UpperCamelCase = processor(text=__a , images=__a) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys()) , ['''pixel_values''', '''input_ids''', '''attention_mask'''])
100
"""simple docstring""" def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) _UpperCamelCase = str(bin(__snake_case ) ) binary_number += "0" * shift_amount return binary_number def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" if number < 0 or shift_amount < 0: raise ValueError('''both inputs must be positive integers''' ) _UpperCamelCase = str(bin(__snake_case ) )[2:] if shift_amount >= len(__snake_case ): return "0b0" _UpperCamelCase = binary_number[: len(__snake_case ) - shift_amount] return "0b" + shifted_binary_number def lowerCamelCase__ ( __snake_case, __snake_case ) -> str: """simple docstring""" if number >= 0: # Get binary representation of positive number _UpperCamelCase = '''0''' + str(bin(__snake_case ) ).strip('''-''' )[2:] else: # Get binary (2's complement) representation of negative number _UpperCamelCase = len(bin(__snake_case )[3:] ) # Find 2's complement of number _UpperCamelCase = bin(abs(__snake_case ) - (1 << binary_number_length) )[3:] _UpperCamelCase = ( '''1''' + '''0''' * (binary_number_length - len(__snake_case )) + binary_number ) if shift_amount >= len(__snake_case ): return "0b" + binary_number[0] * len(__snake_case ) return ( "0b" + binary_number[0] * shift_amount + binary_number[: len(__snake_case ) - shift_amount] ) if __name__ == "__main__": import doctest doctest.testmod()
100
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE_:str = { """configuration_xlm_roberta""": [ """XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLMRobertaConfig""", """XLMRobertaOnnxConfig""", ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:Tuple = ["""XLMRobertaTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:str = ["""XLMRobertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:Optional[int] = [ """XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLMRobertaForCausalLM""", """XLMRobertaForMaskedLM""", """XLMRobertaForMultipleChoice""", """XLMRobertaForQuestionAnswering""", """XLMRobertaForSequenceClassification""", """XLMRobertaForTokenClassification""", """XLMRobertaModel""", """XLMRobertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:str = [ """TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFXLMRobertaForCausalLM""", """TFXLMRobertaForMaskedLM""", """TFXLMRobertaForMultipleChoice""", """TFXLMRobertaForQuestionAnswering""", """TFXLMRobertaForSequenceClassification""", """TFXLMRobertaForTokenClassification""", """TFXLMRobertaModel""", """TFXLMRobertaPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_:List[Any] = [ """FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """FlaxXLMRobertaForMaskedLM""", """FlaxXLMRobertaForCausalLM""", """FlaxXLMRobertaForMultipleChoice""", """FlaxXLMRobertaForQuestionAnswering""", """FlaxXLMRobertaForSequenceClassification""", """FlaxXLMRobertaForTokenClassification""", """FlaxXLMRobertaModel""", """FlaxXLMRobertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_:str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
116
import argparse from pathlib import Path import torch from packaging import version from torch.onnx import export from diffusers import AutoencoderKL SCREAMING_SNAKE_CASE_:List[Any] = version.parse(version.parse(torch.__version__).base_version) < version.parse("""1.11""") def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=False , ) -> int: """simple docstring""" output_path.parent.mkdir(parents=_lowerCAmelCase , exist_ok=_lowerCAmelCase ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( _lowerCAmelCase , _lowerCAmelCase , f=output_path.as_posix() , input_names=_lowerCAmelCase , output_names=_lowerCAmelCase , dynamic_axes=_lowerCAmelCase , do_constant_folding=_lowerCAmelCase , use_external_data_format=_lowerCAmelCase , enable_onnx_checker=_lowerCAmelCase , opset_version=_lowerCAmelCase , ) else: export( _lowerCAmelCase , _lowerCAmelCase , f=output_path.as_posix() , input_names=_lowerCAmelCase , output_names=_lowerCAmelCase , dynamic_axes=_lowerCAmelCase , do_constant_folding=_lowerCAmelCase , opset_version=_lowerCAmelCase , ) @torch.no_grad() def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = False ) -> List[Any]: """simple docstring""" A : Tuple = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): A : Union[str, Any] = """cuda""" elif fpaa and not torch.cuda.is_available(): raise ValueError("""`float16` model export is only supported on GPUs with CUDA""" ) else: A : Any = """cpu""" A : Any = Path(_lowerCAmelCase ) # VAE DECODER A : Union[str, Any] = AutoencoderKL.from_pretrained(model_path + """/vae""" ) A : Any = vae_decoder.config.latent_channels # forward only through the decoder part A : Optional[int] = vae_decoder.decode onnx_export( _lowerCAmelCase , model_args=( torch.randn(1 , _lowerCAmelCase , 25 , 25 ).to(device=_lowerCAmelCase , dtype=_lowerCAmelCase ), False, ) , output_path=output_path / """vae_decoder""" / """model.onnx""" , ordered_input_names=["""latent_sample""", """return_dict"""] , output_names=["""sample"""] , dynamic_axes={ """latent_sample""": {0: """batch""", 1: """channels""", 2: """height""", 3: """width"""}, } , opset=_lowerCAmelCase , ) del vae_decoder if __name__ == "__main__": SCREAMING_SNAKE_CASE_:Tuple = argparse.ArgumentParser() parser.add_argument( """--model_path""", type=str, required=True, help="""Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).""", ) parser.add_argument("""--output_path""", type=str, required=True, help="""Path to the output model.""") parser.add_argument( """--opset""", default=14, type=int, help="""The version of the ONNX operator set to use.""", ) parser.add_argument("""--fp16""", action="""store_true""", default=False, help="""Export the models in `float16` mode""") SCREAMING_SNAKE_CASE_:Tuple = parser.parse_args() print(args.output_path) convert_models(args.model_path, args.output_path, args.opset, args.fpaa) print("""SD: Done: ONNX""")
116
1
# Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' snake_case_ = { 'en': 'Machine learning is great, isn\'t it?', 'ru': 'Машинное обучение - это здорово, не так ли?', 'de': 'Maschinelles Lernen ist großartig, oder?', } # BLUE scores as follows: # "pair": [fairseq, transformers] snake_case_ = { 'ru-en': ['[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)', '39.20'], 'en-ru': ['[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)', '33.47'], 'en-de': ['[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)', '42.83'], 'de-en': ['[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)', '41.35'], } snake_case_ = F'''{src_lang}-{tgt_lang}''' snake_case_ = F''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(UpperCamelCase__ , exist_ok=UpperCamelCase__ ) snake_case_ = os.path.join(UpperCamelCase__ , 'README.md' ) print(F'''Generating {path}''' ) with open(UpperCamelCase__ , 'w' , encoding='utf-8' ) as f: f.write(UpperCamelCase__ ) # make sure we are under the root of the project _UpperCAmelCase : Dict = Path(__file__).resolve().parent.parent.parent _UpperCAmelCase : int = repo_dir / """model_cards""" for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Tuple = model_name.split("""-""") _UpperCAmelCase : Tuple = model_cards_dir / """facebook""" / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
200
def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' return int((input_a, input_a).count(0 ) != 0 ) def __lowerCamelCase ( ): '''simple docstring''' assert nand_gate(0 , 0 ) == 1 assert nand_gate(0 , 1 ) == 1 assert nand_gate(1 , 0 ) == 1 assert nand_gate(1 , 1 ) == 0 if __name__ == "__main__": print(nand_gate(0, 0)) print(nand_gate(0, 1)) print(nand_gate(1, 0)) print(nand_gate(1, 1))
200
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : Optional[int] = logging.get_logger(__name__) __lowerCAmelCase : List[str] = '▁' __lowerCAmelCase : Optional[Any] = {'vocab_file': 'sentencepiece.bpe.model'} __lowerCAmelCase : Optional[int] = { 'vocab_file': { 'facebook/xglm-564M': 'https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model', } } __lowerCAmelCase : Optional[Any] = { 'facebook/xglm-564M': 2048, } class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = VOCAB_FILES_NAMES a__ = PRETRAINED_VOCAB_FILES_MAP a__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ = ["""input_ids""", """attention_mask"""] def __init__( self : Optional[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : str="<s>" , UpperCamelCase__ : Optional[int]="</s>" , UpperCamelCase__ : str="</s>" , UpperCamelCase__ : Optional[Any]="<s>" , UpperCamelCase__ : Optional[int]="<unk>" , UpperCamelCase__ : Optional[int]="<pad>" , UpperCamelCase__ : Optional[Dict[str, Any]] = None , **UpperCamelCase__ : str , ) -> None: """simple docstring""" __magic_name__ = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer __magic_name__ = 7 __magic_name__ = [F'''<madeupword{i}>''' for i in range(self.num_madeup_words )] __magic_name__ = kwargs.get("""additional_special_tokens""" , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) __magic_name__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCamelCase__ ) ) __magic_name__ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __magic_name__ = 1 # Mimic fairseq token-to-id alignment for the first 4 token __magic_name__ = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} __magic_name__ = len(self.sp_model ) __magic_name__ = {F'''<madeupword{i}>''': sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(UpperCamelCase__ ) __magic_name__ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Union[str, Any] ) -> List[str]: """simple docstring""" __magic_name__ = self.__dict__.copy() __magic_name__ = None __magic_name__ = self.sp_model.serialized_model_proto() return state def __setstate__( self : List[str] , UpperCamelCase__ : str ) -> List[Any]: """simple docstring""" __magic_name__ = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __magic_name__ = {} __magic_name__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.sep_token_id] + token_ids_a __magic_name__ = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def _lowercase ( self : Optional[Any] , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None , UpperCamelCase__ : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(UpperCamelCase__ )) return [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] + ([0] * len(UpperCamelCase__ )) def _lowercase ( self : int , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ) -> List[int]: """simple docstring""" __magic_name__ = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def _lowercase ( self : Tuple ) -> List[Any]: """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def _lowercase ( self : Optional[int] ) -> Tuple: """simple docstring""" __magic_name__ = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _lowercase ( self : Any , UpperCamelCase__ : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) def _lowercase ( self : List[Any] , UpperCamelCase__ : Optional[int] ) -> List[Any]: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __magic_name__ = self.sp_model.PieceToId(UpperCamelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _lowercase ( self : str , UpperCamelCase__ : str ) -> Tuple: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _lowercase ( self : Dict , UpperCamelCase__ : Any ) -> Optional[Any]: """simple docstring""" __magic_name__ = """""".join(UpperCamelCase__ ).replace(UpperCamelCase__ , """ """ ).strip() return out_string def _lowercase ( self : Optional[int] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return __magic_name__ = os.path.join( UpperCamelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , """wb""" ) as fi: __magic_name__ = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,)
88
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForPreTraining, AlbertForQuestionAnswering, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertModel, ) from transformers.models.albert.modeling_albert import ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : int=13 , UpperCamelCase__ : Optional[int]=7 , UpperCamelCase__ : Any=True , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : int=True , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : int=99 , UpperCamelCase__ : Any=16 , UpperCamelCase__ : str=36 , UpperCamelCase__ : List[str]=6 , UpperCamelCase__ : List[str]=6 , UpperCamelCase__ : Union[str, Any]=6 , UpperCamelCase__ : int=37 , UpperCamelCase__ : Optional[int]="gelu" , UpperCamelCase__ : List[Any]=0.1 , UpperCamelCase__ : Optional[int]=0.1 , UpperCamelCase__ : int=512 , UpperCamelCase__ : str=16 , UpperCamelCase__ : int=2 , UpperCamelCase__ : List[Any]=0.02 , UpperCamelCase__ : Optional[int]=3 , UpperCamelCase__ : Optional[Any]=4 , UpperCamelCase__ : Dict=None , ) -> Any: """simple docstring""" __magic_name__ = parent __magic_name__ = batch_size __magic_name__ = seq_length __magic_name__ = is_training __magic_name__ = use_input_mask __magic_name__ = use_token_type_ids __magic_name__ = use_labels __magic_name__ = vocab_size __magic_name__ = embedding_size __magic_name__ = hidden_size __magic_name__ = num_hidden_layers __magic_name__ = num_hidden_groups __magic_name__ = num_attention_heads __magic_name__ = intermediate_size __magic_name__ = hidden_act __magic_name__ = hidden_dropout_prob __magic_name__ = attention_probs_dropout_prob __magic_name__ = max_position_embeddings __magic_name__ = type_vocab_size __magic_name__ = type_sequence_label_size __magic_name__ = initializer_range __magic_name__ = num_labels __magic_name__ = num_choices __magic_name__ = scope def _lowercase ( self : Tuple ) -> Dict: """simple docstring""" __magic_name__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __magic_name__ = None if self.use_input_mask: __magic_name__ = random_attention_mask([self.batch_size, self.seq_length] ) __magic_name__ = None if self.use_token_type_ids: __magic_name__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __magic_name__ = None __magic_name__ = None __magic_name__ = None if self.use_labels: __magic_name__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __magic_name__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __magic_name__ = ids_tensor([self.batch_size] , self.num_choices ) __magic_name__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self : Any ) -> List[Any]: """simple docstring""" return AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , num_hidden_groups=self.num_hidden_groups , ) def _lowercase ( self : int , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple , UpperCamelCase__ : int , UpperCamelCase__ : Dict , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any] ) -> Tuple: """simple docstring""" __magic_name__ = AlbertModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) __magic_name__ = model(UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _lowercase ( self : Optional[int] , UpperCamelCase__ : int , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Dict , UpperCamelCase__ : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Union[str, Any] ) -> str: """simple docstring""" __magic_name__ = AlbertForPreTraining(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ , sentence_order_label=UpperCamelCase__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.sop_logits.shape , (self.batch_size, config.num_labels) ) def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : int , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Any , UpperCamelCase__ : List[str] , UpperCamelCase__ : Tuple ) -> Dict: """simple docstring""" __magic_name__ = AlbertForMaskedLM(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str , UpperCamelCase__ : str , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : Tuple ) -> List[Any]: """simple docstring""" __magic_name__ = AlbertForQuestionAnswering(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , start_positions=UpperCamelCase__ , end_positions=UpperCamelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Tuple , UpperCamelCase__ : int ) -> Tuple: """simple docstring""" __magic_name__ = self.num_labels __magic_name__ = AlbertForSequenceClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowercase ( self : int , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Tuple , UpperCamelCase__ : str , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : Optional[int] ) -> int: """simple docstring""" __magic_name__ = self.num_labels __magic_name__ = AlbertForTokenClassification(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self : Optional[int] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : List[str] , UpperCamelCase__ : int , UpperCamelCase__ : str ) -> List[Any]: """simple docstring""" __magic_name__ = self.num_choices __magic_name__ = AlbertForMultipleChoice(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __magic_name__ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __magic_name__ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __magic_name__ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __magic_name__ = model( UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _lowercase ( self : int ) -> Optional[int]: """simple docstring""" __magic_name__ = self.prepare_config_and_inputs() ( ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ( __magic_name__ ) , ) = config_and_inputs __magic_name__ = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _A , _A , unittest.TestCase ): '''simple docstring''' a__ = ( ( AlbertModel, AlbertForPreTraining, AlbertForMaskedLM, AlbertForMultipleChoice, AlbertForSequenceClassification, AlbertForTokenClassification, AlbertForQuestionAnswering, ) if is_torch_available() else () ) a__ = ( { """feature-extraction""": AlbertModel, """fill-mask""": AlbertForMaskedLM, """question-answering""": AlbertForQuestionAnswering, """text-classification""": AlbertForSequenceClassification, """token-classification""": AlbertForTokenClassification, """zero-shot""": AlbertForSequenceClassification, } if is_torch_available() else {} ) a__ = True def _lowercase ( self : str , UpperCamelCase__ : Any , UpperCamelCase__ : Dict , UpperCamelCase__ : Union[str, Any]=False ) -> Union[str, Any]: """simple docstring""" __magic_name__ = super()._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ , return_labels=UpperCamelCase__ ) if return_labels: if model_class in get_values(UpperCamelCase__ ): __magic_name__ = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=UpperCamelCase__ ) __magic_name__ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCamelCase__ ) return inputs_dict def _lowercase ( self : int ) -> int: """simple docstring""" __magic_name__ = AlbertModelTester(self ) __magic_name__ = ConfigTester(self , config_class=UpperCamelCase__ , hidden_size=37 ) def _lowercase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" self.config_tester.run_common_tests() def _lowercase ( self : Dict ) -> Dict: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def _lowercase ( self : int ) -> List[str]: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*UpperCamelCase__ ) def _lowercase ( self : List[Any] ) -> Any: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCamelCase__ ) def _lowercase ( self : Dict ) -> Tuple: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*UpperCamelCase__ ) def _lowercase ( self : Dict ) -> List[Any]: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCamelCase__ ) def _lowercase ( self : Union[str, Any] ) -> Any: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCamelCase__ ) def _lowercase ( self : Tuple ) -> Optional[Any]: """simple docstring""" __magic_name__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __magic_name__ = type self.model_tester.create_and_check_model(*UpperCamelCase__ ) @slow def _lowercase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" for model_name in ALBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __magic_name__ = AlbertModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) @require_torch class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def _lowercase ( self : Dict ) -> Union[str, Any]: """simple docstring""" __magic_name__ = AlbertModel.from_pretrained("""albert-base-v2""" ) __magic_name__ = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __magic_name__ = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __magic_name__ = model(UpperCamelCase__ , attention_mask=UpperCamelCase__ )[0] __magic_name__ = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , UpperCamelCase__ ) __magic_name__ = torch.tensor( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , UpperCamelCase__ , atol=1E-4 ) )
88
1
from __future__ import annotations SCREAMING_SNAKE_CASE__ = tuple[int, int, int] SCREAMING_SNAKE_CASE__ = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase SCREAMING_SNAKE_CASE__ = "ABCDEFGHIJKLMNOPQRSTUVWXYZ" # -------------------------- default selection -------------------------- # rotors -------------------------- SCREAMING_SNAKE_CASE__ = "EGZWVONAHDCLFQMSIPJBYUKXTR" SCREAMING_SNAKE_CASE__ = "FOBHMDKEXQNRAULPGSJVTYICZW" SCREAMING_SNAKE_CASE__ = "ZJXESIUQLHAVRMDOYGTNFWPBKC" # reflector -------------------------- SCREAMING_SNAKE_CASE__ = { "A": "N", "N": "A", "B": "O", "O": "B", "C": "P", "P": "C", "D": "Q", "Q": "D", "E": "R", "R": "E", "F": "S", "S": "F", "G": "T", "T": "G", "H": "U", "U": "H", "I": "V", "V": "I", "J": "W", "W": "J", "K": "X", "X": "K", "L": "Y", "Y": "L", "M": "Z", "Z": "M", } # -------------------------- extra rotors -------------------------- SCREAMING_SNAKE_CASE__ = "RMDJXFUWGISLHVTCQNKYPBEZOA" SCREAMING_SNAKE_CASE__ = "SGLCPQWZHKXAREONTFBVIYJUDM" SCREAMING_SNAKE_CASE__ = "HVSICLTYKQUBXDWAJZOMFGPREN" SCREAMING_SNAKE_CASE__ = "RZWQHFMVDBKICJLNTUXAGYPSOE" SCREAMING_SNAKE_CASE__ = "LFKIJODBEGAMQPXVUHYSTCZRWN" SCREAMING_SNAKE_CASE__ = "KOAEGVDHXPQZMLFTYWJNBRCIUS" def __magic_name__ ( __lowerCAmelCase : RotorPositionT , __lowerCAmelCase : RotorSelectionT , __lowerCAmelCase : str ) -> int: if (unique_rotsel := len(set(lowerCamelCase__ ) )) < 3: __lowerCamelCase = f'''Please use 3 unique rotors (not {unique_rotsel})''' raise Exception(lowerCamelCase__ ) # Checks if rotor positions are valid __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = rotpos if not 0 < rotorposa <= len(lowerCamelCase__ ): __lowerCamelCase = f'''First rotor position is not within range of 1..26 ({rotorposa}''' raise ValueError(lowerCamelCase__ ) if not 0 < rotorposa <= len(lowerCamelCase__ ): __lowerCamelCase = f'''Second rotor position is not within range of 1..26 ({rotorposa})''' raise ValueError(lowerCamelCase__ ) if not 0 < rotorposa <= len(lowerCamelCase__ ): __lowerCamelCase = f'''Third rotor position is not within range of 1..26 ({rotorposa})''' raise ValueError(lowerCamelCase__ ) # Validates string and returns dict __lowerCamelCase = _plugboard(lowerCamelCase__ ) return rotpos, rotsel, pbdict def __magic_name__ ( __lowerCAmelCase : str ) -> Any: if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): __lowerCamelCase = f'''Plugboard setting isn\'t type string ({type(lowerCamelCase__ )})''' raise TypeError(lowerCamelCase__ ) elif len(lowerCamelCase__ ) % 2 != 0: __lowerCamelCase = f'''Odd number of symbols ({len(lowerCamelCase__ )})''' raise Exception(lowerCamelCase__ ) elif pbstring == "": return {} pbstring.replace(''' ''' , '''''' ) # Checks if all characters are unique __lowerCamelCase = set() for i in pbstring: if i not in abc: __lowerCamelCase = f'''\'{i}\' not in list of symbols''' raise Exception(lowerCamelCase__ ) elif i in tmppbl: __lowerCamelCase = f'''Duplicate symbol ({i})''' raise Exception(lowerCamelCase__ ) else: tmppbl.add(lowerCamelCase__ ) del tmppbl # Created the dictionary __lowerCamelCase = {} for j in range(0 , len(lowerCamelCase__ ) - 1 , 2 ): __lowerCamelCase = pbstring[j + 1] __lowerCamelCase = pbstring[j] return pb def __magic_name__ ( __lowerCAmelCase : str , __lowerCAmelCase : RotorPositionT , __lowerCAmelCase : RotorSelectionT = (rotora, rotora, rotora) , __lowerCAmelCase : str = "" , ) -> Tuple: __lowerCamelCase = text.upper() __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = _validator( lowerCamelCase__ , lowerCamelCase__ , plugb.upper() ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = rotor_position __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 __lowerCamelCase = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: __lowerCamelCase = plugboard[symbol] # rotor ra -------------------------- __lowerCamelCase = abc.index(lowerCamelCase__ ) + rotorposa __lowerCamelCase = rotora[index % len(lowerCamelCase__ )] # rotor rb -------------------------- __lowerCamelCase = abc.index(lowerCamelCase__ ) + rotorposa __lowerCamelCase = rotora[index % len(lowerCamelCase__ )] # rotor rc -------------------------- __lowerCamelCase = abc.index(lowerCamelCase__ ) + rotorposa __lowerCamelCase = rotora[index % len(lowerCamelCase__ )] # reflector -------------------------- # this is the reason you don't need another machine to decipher __lowerCamelCase = reflector[symbol] # 2nd rotors __lowerCamelCase = abc[rotora.index(lowerCamelCase__ ) - rotorposa] __lowerCamelCase = abc[rotora.index(lowerCamelCase__ ) - rotorposa] __lowerCamelCase = abc[rotora.index(lowerCamelCase__ ) - rotorposa] # 2nd plugboard if symbol in plugboard: __lowerCamelCase = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(lowerCamelCase__ ): __lowerCamelCase = 0 rotorposa += 1 if rotorposa >= len(lowerCamelCase__ ): __lowerCamelCase = 0 rotorposa += 1 if rotorposa >= len(lowerCamelCase__ ): __lowerCamelCase = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(lowerCamelCase__ ) return "".join(lowerCamelCase__ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = "This is my Python script that emulates the Enigma machine from WWII." SCREAMING_SNAKE_CASE__ = (1, 1, 1) SCREAMING_SNAKE_CASE__ = "pictures" SCREAMING_SNAKE_CASE__ = (rotora, rotora, rotora) SCREAMING_SNAKE_CASE__ = enigma(message, rotor_pos, rotor_sel, pb) print("Encrypted message:", en) print("Decrypted message:", enigma(en, rotor_pos, rotor_sel, pb))
350
def __magic_name__ ( __lowerCAmelCase : int , __lowerCAmelCase : int ) -> int: return abs(__lowerCAmelCase ) if a == 0 else greatest_common_divisor(b % a , __lowerCAmelCase ) def __magic_name__ ( __lowerCAmelCase : int , __lowerCAmelCase : int ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. __lowerCamelCase , __lowerCamelCase = y, x % y return abs(__lowerCAmelCase ) def __magic_name__ ( ) -> Tuple: try: __lowerCamelCase = input('''Enter two integers separated by comma (,): ''' ).split(''',''' ) __lowerCamelCase = int(nums[0] ) __lowerCamelCase = int(nums[1] ) print( f'''greatest_common_divisor({num_a}, {num_a}) = ''' f'''{greatest_common_divisor(__lowerCAmelCase , __lowerCAmelCase )}''' ) print(f'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(__lowerCAmelCase , __lowerCAmelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print('''Wrong input''' ) if __name__ == "__main__": main()
339
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase: str = logging.get_logger(__name__) class a__( lowerCamelCase__ ): lowercase__ = ["""pixel_values"""] def __init__( self : Union[str, Any] , __snake_case : Any = True , __snake_case : List[Any] = None , __snake_case : Union[str, Any] = PILImageResampling.BICUBIC , __snake_case : List[str] = True , __snake_case : int = 1 / 2_55 , __snake_case : int = True , __snake_case : int = None , __snake_case : Optional[Any] = None , __snake_case : List[Any] = True , **__snake_case : Tuple , ): super().__init__(**UpperCamelCase_ ) a : Dict = size if size is not None else {'''height''': 3_84, '''width''': 3_84} a : int = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) a : Tuple = do_resize a : List[str] = size a : Dict = resample a : Optional[Any] = do_rescale a : List[str] = rescale_factor a : Any = do_normalize a : List[str] = image_mean if image_mean is not None else OPENAI_CLIP_MEAN a : str = image_std if image_std is not None else OPENAI_CLIP_STD a : Any = do_convert_rgb def lowercase_ ( self : List[Any] , __snake_case : Any , __snake_case : List[str] , __snake_case : int = PILImageResampling.BICUBIC , __snake_case : Tuple = None , **__snake_case : Optional[int] , ): a : Dict = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}""" ) a : Union[str, Any] = (size['''height'''], size['''width''']) return resize(UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowercase_ ( self : str , __snake_case : Any , __snake_case : List[str] , __snake_case : Union[str, Any] = None , **__snake_case : Dict , ): return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowercase_ ( self : List[Any] , __snake_case : str , __snake_case : int , __snake_case : Optional[int] , __snake_case : Optional[int] = None , **__snake_case : Tuple , ): return normalize(UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowercase_ ( self : List[str] , __snake_case : Optional[Any] , __snake_case : Optional[int] = None , __snake_case : List[Any] = None , __snake_case : Optional[Any] = None , __snake_case : Optional[Any] = None , __snake_case : Optional[Any] = None , __snake_case : List[Any] = None , __snake_case : Dict = None , __snake_case : Optional[Any] = None , __snake_case : Optional[Any] = None , __snake_case : str = None , __snake_case : Tuple = ChannelDimension.FIRST , **__snake_case : List[Any] , ): a : int = do_resize if do_resize is not None else self.do_resize a : Union[str, Any] = resample if resample is not None else self.resample a : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale a : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor a : Optional[Any] = do_normalize if do_normalize is not None else self.do_normalize a : Union[str, Any] = image_mean if image_mean is not None else self.image_mean a : Any = image_std if image_std is not None else self.image_std a : str = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb a : List[Any] = size if size is not None else self.size a : List[str] = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) a : str = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: a : List[Any] = [convert_to_rgb(UpperCamelCase_ ) for image in images] # All transformations expect numpy arrays. a : Optional[int] = [to_numpy_array(UpperCamelCase_ ) for image in images] if do_resize: a : Any = [self.resize(image=UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ ) for image in images] if do_rescale: a : List[Any] = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_normalize: a : Any = [self.normalize(image=UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ ) for image in images] a : List[Any] = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] a : List[Any] = BatchFeature(data={'pixel_values': images} , tensor_type=UpperCamelCase_ ) return encoded_outputs
297
"""simple docstring""" def __UpperCAmelCase ( __UpperCamelCase ): __lowercase : str = [1] __lowercase ,__lowercase ,__lowercase : List[str] = 0, 0, 0 __lowercase : List[str] = ugly_nums[ia] * 2 __lowercase : Any = ugly_nums[ia] * 3 __lowercase : str = ugly_nums[ia] * 5 for _ in range(1 , __UpperCamelCase ): __lowercase : Union[str, Any] = min(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) ugly_nums.append(__UpperCamelCase ) if next_num == next_a: ia += 1 __lowercase : List[str] = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 __lowercase : int = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 __lowercase : Optional[int] = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(F"{ugly_numbers(2_0_0) = }")
249
0
import sys from collections import defaultdict class A : '''simple docstring''' def __init__( self : Union[str, Any] ) -> int: """simple docstring""" A__ = [] def a_ ( self : Union[str, Any] , __lowerCAmelCase : Optional[Any] ) -> List[Any]: """simple docstring""" return self.node_position[vertex] def a_ ( self : str , __lowerCAmelCase : Any , __lowerCAmelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" A__ = pos def a_ ( self : List[str] , __lowerCAmelCase : Any , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple ) -> int: """simple docstring""" if start > size // 2 - 1: return else: if 2 * start + 2 >= size: A__ = 2 * start + 1 else: if heap[2 * start + 1] < heap[2 * start + 2]: A__ = 2 * start + 1 else: A__ = 2 * start + 2 if heap[smallest_child] < heap[start]: A__ , A__ = heap[smallest_child], positions[smallest_child] A__ , A__ = ( heap[start], positions[start], ) A__ , A__ = temp, tempa A__ = self.get_position(positions[smallest_child] ) self.set_position( positions[smallest_child] , self.get_position(positions[start] ) ) self.set_position(positions[start] , __lowerCAmelCase ) self.top_to_bottom(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def a_ ( self : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Dict ) -> Tuple: """simple docstring""" A__ = position[index] while index != 0: A__ = int((index - 2) / 2 ) if index % 2 == 0 else int((index - 1) / 2 ) if val < heap[parent]: A__ = heap[parent] A__ = position[parent] self.set_position(position[parent] , __lowerCAmelCase ) else: A__ = val A__ = temp self.set_position(__lowerCAmelCase , __lowerCAmelCase ) break A__ = parent else: A__ = val A__ = temp self.set_position(__lowerCAmelCase , 0 ) def a_ ( self : str , __lowerCAmelCase : str , __lowerCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" A__ = len(__lowerCAmelCase ) // 2 - 1 for i in range(__lowerCAmelCase , -1 , -1 ): self.top_to_bottom(__lowerCAmelCase , __lowerCAmelCase , len(__lowerCAmelCase ) , __lowerCAmelCase ) def a_ ( self : Union[str, Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[int] ) -> Union[str, Any]: """simple docstring""" A__ = positions[0] A__ = sys.maxsize self.top_to_bottom(__lowerCAmelCase , 0 , len(__lowerCAmelCase ) , __lowerCAmelCase ) return temp def __lowerCamelCase ( __a :List[Any] ) -> List[Any]: """simple docstring""" A__ = Heap() A__ = [0] * len(__a ) A__ = [-1] * len(__a ) # Neighboring Tree Vertex of selected vertex # Minimum Distance of explored vertex with neighboring vertex of partial tree # formed in graph A__ = [] # Heap of Distance of vertices from their neighboring vertex A__ = [] for vertex in range(len(__a ) ): distance_tv.append(sys.maxsize ) positions.append(__a ) heap.node_position.append(__a ) A__ = [] A__ = 1 A__ = sys.maxsize for neighbor, distance in adjacency_list[0]: A__ = 0 A__ = distance heap.heapify(__a , __a ) for _ in range(1 , len(__a ) ): A__ = heap.delete_minimum(__a , __a ) if visited[vertex] == 0: tree_edges.append((nbr_tv[vertex], vertex) ) A__ = 1 for neighbor, distance in adjacency_list[vertex]: if ( visited[neighbor] == 0 and distance < distance_tv[heap.get_position(__a )] ): A__ = distance heap.bottom_to_top( __a , heap.get_position(__a ) , __a , __a ) A__ = vertex return tree_edges if __name__ == "__main__": # pragma: no cover # < --------- Prims Algorithm --------- > A : Optional[Any] = int(input('''Enter number of edges: ''').strip()) A : Tuple = defaultdict(list) for _ in range(edges_number): A : int = [int(x) for x in input().strip().split()] adjacency_list[edge[0]].append([edge[1], edge[2]]) adjacency_list[edge[1]].append([edge[0], edge[2]]) print(prisms_algorithm(adjacency_list))
276
from typing import TYPE_CHECKING from ...utils import _LazyModule A : Optional[Any] = {'''processing_wav2vec2_with_lm''': ['''Wav2Vec2ProcessorWithLM''']} if TYPE_CHECKING: from .processing_wavaveca_with_lm import WavaVecaProcessorWithLM else: import sys A : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
276
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __snake_case = logging.get_logger(__name__) __snake_case = '''▁''' __snake_case = {'''vocab_file''': '''sentencepiece.bpe.model'''} __snake_case = { '''vocab_file''': { '''facebook/mbart-large-en-ro''': ( '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model''' ), '''facebook/mbart-large-cc25''': ( '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model''' ), } } __snake_case = { '''facebook/mbart-large-en-ro''': 1024, '''facebook/mbart-large-cc25''': 1024, } # fmt: off __snake_case = ['''ar_AR''', '''cs_CZ''', '''de_DE''', '''en_XX''', '''es_XX''', '''et_EE''', '''fi_FI''', '''fr_XX''', '''gu_IN''', '''hi_IN''', '''it_IT''', '''ja_XX''', '''kk_KZ''', '''ko_KR''', '''lt_LT''', '''lv_LV''', '''my_MM''', '''ne_NP''', '''nl_XX''', '''ro_RO''', '''ru_RU''', '''si_LK''', '''tr_TR''', '''vi_VN''', '''zh_CN'''] class lowercase ( A__ ): """simple docstring""" _a = VOCAB_FILES_NAMES _a = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _a = PRETRAINED_VOCAB_FILES_MAP _a = ['input_ids', 'attention_mask'] _a = [] _a = [] def __init__( self , UpperCamelCase_ , UpperCamelCase_="<s>" , UpperCamelCase_="</s>" , UpperCamelCase_="</s>" , UpperCamelCase_="<s>" , UpperCamelCase_="<unk>" , UpperCamelCase_="<pad>" , UpperCamelCase_="<mask>" , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_=None , UpperCamelCase_ = None , UpperCamelCase_=None , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :Dict = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else mask_token UpperCamelCase__ :int = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , src_lang=UpperCamelCase_ , tgt_lang=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase_ , ) UpperCamelCase__ :int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCamelCase_ ) ) UpperCamelCase__ :Optional[int] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token UpperCamelCase__ :Dict = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCamelCase__ :Tuple = 1 UpperCamelCase__ :int = len(self.sp_model ) UpperCamelCase__ :Dict = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(UpperCamelCase_ ) } UpperCamelCase__ :List[Any] = {v: k for k, v in self.lang_code_to_id.items()} UpperCamelCase__ :Any = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) UpperCamelCase__ :Union[str, Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} UpperCamelCase__ :Union[str, Any] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) UpperCamelCase__ :Any = src_lang if src_lang is not None else '''en_XX''' UpperCamelCase__ :Optional[Any] = self.lang_code_to_id[self._src_lang] UpperCamelCase__ :Union[str, Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ): '''simple docstring''' UpperCamelCase__ :Dict = self.__dict__.copy() UpperCamelCase__ :int = None UpperCamelCase__ :Dict = self.sp_model.serialized_model_proto() return state def __setstate__( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Tuple = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCamelCase__ :Optional[int] = {} UpperCamelCase__ :Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def lowerCAmelCase__ ( self ): '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def lowerCAmelCase__ ( self ): '''simple docstring''' return self._src_lang @src_lang.setter def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None , UpperCamelCase_ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_ ) UpperCamelCase__ :List[str] = [1] * len(self.prefix_tokens ) UpperCamelCase__ :int = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(UpperCamelCase_ )) + suffix_ones return prefix_ones + ([0] * len(UpperCamelCase_ )) + ([0] * len(UpperCamelCase_ )) + suffix_ones def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' UpperCamelCase__ :Optional[int] = [self.sep_token_id] UpperCamelCase__ :List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) UpperCamelCase__ :Tuple = src_lang UpperCamelCase__ :Optional[Any] = self(UpperCamelCase_ , add_special_tokens=UpperCamelCase_ , return_tensors=UpperCamelCase_ , **UpperCamelCase_ ) UpperCamelCase__ :List[str] = self.convert_tokens_to_ids(UpperCamelCase_ ) UpperCamelCase__ :Dict = tgt_lang_id return inputs def lowerCAmelCase__ ( self ): '''simple docstring''' UpperCamelCase__ :Tuple = {self.convert_ids_to_tokens(UpperCamelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCamelCase__ :Any = self.sp_model.PieceToId(UpperCamelCase_ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :List[str] = ''''''.join(UpperCamelCase_ ).replace(UpperCamelCase_ , ''' ''' ).strip() return out_string def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = None ): '''simple docstring''' if not os.path.isdir(UpperCamelCase_ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCamelCase__ :int = os.path.join( UpperCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase_ , '''wb''' ) as fi: UpperCamelCase__ :Any = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_ ) return (out_vocab_file,) def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ = "en_XX" , UpperCamelCase_ = None , UpperCamelCase_ = "ro_RO" , **UpperCamelCase_ , ): '''simple docstring''' UpperCamelCase__ :Optional[Any] = src_lang UpperCamelCase__ :Optional[Any] = tgt_lang return super().prepare_seqaseq_batch(UpperCamelCase_ , UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def lowerCAmelCase__ ( self ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Any = self.lang_code_to_id[src_lang] UpperCamelCase__ :int = [] UpperCamelCase__ :Union[str, Any] = [self.eos_token_id, self.cur_lang_code] def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' UpperCamelCase__ :Dict = self.lang_code_to_id[lang] UpperCamelCase__ :Optional[Any] = [] UpperCamelCase__ :Tuple = [self.eos_token_id, self.cur_lang_code]
97
'''simple docstring''' from pathlib import Path import fire from tqdm import tqdm def a ( __a="ro" , __a="en" , __a="wmt16" , __a=None ) -> None: '''simple docstring''' try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError('''run pip install datasets''' ) UpperCamelCase__ :int = f'''{src_lang}-{tgt_lang}''' print(f'''Converting {dataset}-{pair}''' ) UpperCamelCase__ :Tuple = datasets.load_dataset(__a , __a ) if save_dir is None: UpperCamelCase__ :Any = f'''{dataset}-{pair}''' UpperCamelCase__ :Dict = Path(__a ) save_dir.mkdir(exist_ok=__a ) for split in ds.keys(): print(f'''Splitting {split} with {ds[split].num_rows} records''' ) # to save to val.source, val.target like summary datasets UpperCamelCase__ :Dict = '''val''' if split == '''validation''' else split UpperCamelCase__ :List[Any] = save_dir.joinpath(f'''{fn}.source''' ) UpperCamelCase__ :int = save_dir.joinpath(f'''{fn}.target''' ) UpperCamelCase__ :Union[str, Any] = src_path.open('''w+''' ) UpperCamelCase__ :Tuple = tgt_path.open('''w+''' ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): UpperCamelCase__ :Union[str, Any] = x['''translation'''] src_fp.write(ex[src_lang] + '''\n''' ) tgt_fp.write(ex[tgt_lang] + '''\n''' ) print(f'''Saved {dataset} dataset to {save_dir}''' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
97
1
def _UpperCamelCase ( UpperCamelCase_ : int ) -> list: """simple docstring""" lowerCAmelCase__ = int(UpperCamelCase_ ) if n_element < 1: lowerCAmelCase__ = ValueError('a should be a positive number' ) raise my_error lowerCAmelCase__ = [1] lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = (0, 0, 0) lowerCAmelCase__ = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": __snake_case : List[Any] = input("""Enter the last number (nth term) of the Hamming Number Series: """) print("""Formula of Hamming Number Series => 2^i * 3^j * 5^k""") __snake_case : Union[str, Any] = hamming(int(n)) print("""-----------------------------------------------------""") print(f'The list with nth numbers is: {hamming_numbers}') print("""-----------------------------------------------------""")
355
import sacrebleu as scb from packaging import version from sacrebleu import TER import datasets __snake_case : Optional[int] = """\ @inproceedings{snover-etal-2006-study, title = \"A Study of Translation Edit Rate with Targeted Human Annotation\", author = \"Snover, Matthew and Dorr, Bonnie and Schwartz, Rich and Micciulla, Linnea and Makhoul, John\", booktitle = \"Proceedings of the 7th Conference of the Association for Machine Translation in the Americas: Technical Papers\", month = aug # \" 8-12\", year = \"2006\", address = \"Cambridge, Massachusetts, USA\", publisher = \"Association for Machine Translation in the Americas\", url = \"https://aclanthology.org/2006.amta-papers.25\", pages = \"223--231\", } @inproceedings{post-2018-call, title = \"A Call for Clarity in Reporting {BLEU} Scores\", author = \"Post, Matt\", booktitle = \"Proceedings of the Third Conference on Machine Translation: Research Papers\", month = oct, year = \"2018\", address = \"Belgium, Brussels\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/W18-6319\", pages = \"186--191\", } """ __snake_case : List[str] = """\ TER (Translation Edit Rate, also called Translation Error Rate) is a metric to quantify the edit operations that a hypothesis requires to match a reference translation. We use the implementation that is already present in sacrebleu (https://github.com/mjpost/sacreBLEU#ter), which in turn is inspired by the TERCOM implementation, which can be found here: https://github.com/jhclark/tercom. The implementation here is slightly different from sacrebleu in terms of the required input format. The length of the references and hypotheses lists need to be the same, so you may need to transpose your references compared to sacrebleu's required input format. See https://github.com/huggingface/datasets/issues/3154#issuecomment-950746534 See the README.md file at https://github.com/mjpost/sacreBLEU#ter for more information. """ __snake_case : Dict = """ Produces TER scores alongside the number of edits and reference length. Args: predictions (list of str): The system stream (a sequence of segments). references (list of list of str): A list of one or more reference streams (each a sequence of segments). normalized (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. ignore_punct (boolean): If `True`, applies basic tokenization and normalization to sentences. Defaults to `False`. support_zh_ja_chars (boolean): If `True`, tokenization/normalization supports processing of Chinese characters, as well as Japanese Kanji, Hiragana, Katakana, and Phonetic Extensions of Katakana. Only applies if `normalized = True`. Defaults to `False`. case_sensitive (boolean): If `False`, makes all predictions and references lowercase to ignore differences in case. Defaults to `False`. Returns: 'score' (float): TER score (num_edits / sum_ref_lengths * 100) 'num_edits' (int): The cumulative number of edits 'ref_length' (float): The cumulative average reference length Examples: Example 1: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 150.0, 'num_edits': 15, 'ref_length': 10.0} Example 2: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... case_sensitive=True) >>> print(results) {'score': 62.5, 'num_edits': 5, 'ref_length': 8.0} Example 3: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... normalized=True, ... case_sensitive=True) >>> print(results) {'score': 57.14285714285714, 'num_edits': 6, 'ref_length': 10.5} Example 4: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 0.0, 'num_edits': 0, 'ref_length': 8.0} Example 5: >>> predictions = [\"does this sentence match??\", ... \"what about this sentence?\", ... \"What did the TER metric user say to the developer?\"] >>> references = [[\"does this sentence match\", \"does this sentence match!?!\"], ... [\"wHaT aBoUt ThIs SeNtEnCe?\", \"wHaT aBoUt ThIs SeNtEnCe?\"], ... [\"Your jokes are...\", \"...TERrible\"]] >>> ter = datasets.load_metric(\"ter\") >>> results = ter.compute(predictions=predictions, ... references=references, ... ignore_punct=True, ... case_sensitive=False) >>> print(results) {'score': 100.0, 'num_edits': 10, 'ref_length': 10.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class __SCREAMING_SNAKE_CASE ( datasets.Metric): def UpperCamelCase__ ( self ): """simple docstring""" if version.parse(scb.__version__ ) < version.parse('1.4.12' ): raise ImportWarning( 'To use `sacrebleu`, the module `sacrebleu>=1.4.12` is required, and the current version of `sacrebleu` doesn\'t match this condition.\n' 'You can install it with `pip install "sacrebleu>=1.4.12"`.' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='http://www.cs.umd.edu/~snover/tercom/' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Sequence(datasets.Value('string' , id='sequence' ) , id='references' ), } ) , codebase_urls=['https://github.com/mjpost/sacreBLEU#ter'] , reference_urls=[ 'https://github.com/jhclark/tercom', ] , ) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = False , _UpperCamelCase = False , _UpperCamelCase = False , _UpperCamelCase = False , ): """simple docstring""" lowerCAmelCase__ = len(references[0] ) if any(len(_UpperCamelCase ) != references_per_prediction for refs in references ): raise ValueError('Sacrebleu requires the same number of references for each prediction' ) lowerCAmelCase__ = [[refs[i] for refs in references] for i in range(_UpperCamelCase )] lowerCAmelCase__ = TER( normalized=_UpperCamelCase , no_punct=_UpperCamelCase , asian_support=_UpperCamelCase , case_sensitive=_UpperCamelCase , ) lowerCAmelCase__ = sb_ter.corpus_score(_UpperCamelCase , _UpperCamelCase ) return {"score": output.score, "num_edits": output.num_edits, "ref_length": output.ref_length}
122
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def lowerCamelCase_ ( _UpperCamelCase ) -> str: """simple docstring""" snake_case_ : int = botoa.client('''iam''' ) snake_case_ : Optional[int] = { """Version""": """2012-10-17""", """Statement""": [ {"""Effect""": """Allow""", """Principal""": {"""Service""": """sagemaker.amazonaws.com"""}, """Action""": """sts:AssumeRole"""} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=a__ , AssumeRolePolicyDocument=json.dumps(a__ , indent=2 ) ) snake_case_ : str = { """Version""": """2012-10-17""", """Statement""": [ { """Effect""": """Allow""", """Action""": [ """sagemaker:*""", """ecr:GetDownloadUrlForLayer""", """ecr:BatchGetImage""", """ecr:BatchCheckLayerAvailability""", """ecr:GetAuthorizationToken""", """cloudwatch:PutMetricData""", """cloudwatch:GetMetricData""", """cloudwatch:GetMetricStatistics""", """cloudwatch:ListMetrics""", """logs:CreateLogGroup""", """logs:CreateLogStream""", """logs:DescribeLogStreams""", """logs:PutLogEvents""", """logs:GetLogEvents""", """s3:CreateBucket""", """s3:ListBucket""", """s3:GetBucketLocation""", """s3:GetObject""", """s3:PutObject""", ], """Resource""": """*""", } ], } # attach policy to role iam_client.put_role_policy( RoleName=a__ , PolicyName=f'''{role_name}_policy_permission''' , PolicyDocument=json.dumps(a__ , indent=2 ) , ) except iam_client.exceptions.EntityAlreadyExistsException: print(f'''role {role_name} already exists. Using existing one''' ) def lowerCamelCase_ ( _UpperCamelCase ) -> Union[str, Any]: """simple docstring""" snake_case_ : str = botoa.client('''iam''' ) return iam_client.get_role(RoleName=a__ )["Role"]["Arn"] def lowerCamelCase_ ( ) -> Tuple: """simple docstring""" snake_case_ : Tuple = _ask_options( '''How do you want to authorize?''' , ['''AWS Profile''', '''Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) '''] , a__ , ) snake_case_ : Dict = None if credentials_configuration == 0: snake_case_ : Dict = _ask_field('''Enter your AWS Profile name: [default] ''' , default='''default''' ) snake_case_ : int = aws_profile else: print( '''Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,''' '''`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`''' ) snake_case_ : Optional[Any] = _ask_field('''AWS Access Key ID: ''' ) snake_case_ : Union[str, Any] = aws_access_key_id snake_case_ : Dict = _ask_field('''AWS Secret Access Key: ''' ) snake_case_ : Tuple = aws_secret_access_key snake_case_ : Union[str, Any] = _ask_field('''Enter your AWS Region: [us-east-1]''' , default='''us-east-1''' ) snake_case_ : List[Any] = aws_region snake_case_ : str = _ask_options( '''Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?''' , ['''Provide IAM Role name''', '''Create new IAM role using credentials'''] , a__ , ) if role_management == 0: snake_case_ : str = _ask_field('''Enter your IAM role name: ''' ) else: snake_case_ : Optional[int] = """accelerate_sagemaker_execution_role""" print(f'''Accelerate will create an iam role "{iam_role_name}" using the provided credentials''' ) _create_iam_role_for_sagemaker(a__ ) snake_case_ : Union[str, Any] = _ask_field( '''Do you want to use custom Docker image? [yes/NO]: ''' , _convert_yes_no_to_bool , default=a__ , error_message='''Please enter yes or no.''' , ) snake_case_ : List[str] = None if is_custom_docker_image: snake_case_ : Union[str, Any] = _ask_field('''Enter your Docker image: ''' , lambda _UpperCamelCase : str(a__ ).lower() ) snake_case_ : List[str] = _ask_field( '''Do you want to provide SageMaker input channels with data locations? [yes/NO]: ''' , _convert_yes_no_to_bool , default=a__ , error_message='''Please enter yes or no.''' , ) snake_case_ : str = None if is_sagemaker_inputs_enabled: snake_case_ : Any = _ask_field( '''Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): ''' , lambda _UpperCamelCase : str(a__ ).lower() , ) snake_case_ : Optional[Any] = _ask_field( '''Do you want to enable SageMaker metrics? [yes/NO]: ''' , _convert_yes_no_to_bool , default=a__ , error_message='''Please enter yes or no.''' , ) snake_case_ : List[str] = None if is_sagemaker_metrics_enabled: snake_case_ : Tuple = _ask_field( '''Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): ''' , lambda _UpperCamelCase : str(a__ ).lower() , ) snake_case_ : Dict = _ask_options( '''What is the distributed mode?''' , ['''No distributed training''', '''Data parallelism'''] , _convert_sagemaker_distributed_mode , ) snake_case_ : int = {} snake_case_ : Union[str, Any] = _ask_field( '''Do you wish to optimize your script with torch dynamo?[yes/NO]:''' , _convert_yes_no_to_bool , default=a__ , error_message='''Please enter yes or no.''' , ) if use_dynamo: snake_case_ : Any = """dynamo_""" snake_case_ : Union[str, Any] = _ask_options( '''Which dynamo backend would you like to use?''' , [x.lower() for x in DYNAMO_BACKENDS] , _convert_dynamo_backend , default=2 , ) snake_case_ : Optional[Any] = _ask_field( '''Do you want to customize the defaults sent to torch.compile? [yes/NO]: ''' , _convert_yes_no_to_bool , default=a__ , error_message='''Please enter yes or no.''' , ) if use_custom_options: snake_case_ : str = _ask_options( '''Which mode do you want to use?''' , a__ , lambda _UpperCamelCase : TORCH_DYNAMO_MODES[int(a__ )] , default='''default''' , ) snake_case_ : int = _ask_field( '''Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: ''' , _convert_yes_no_to_bool , default=a__ , error_message='''Please enter yes or no.''' , ) snake_case_ : Optional[Any] = _ask_field( '''Do you want to enable dynamic shape tracing? [yes/NO]: ''' , _convert_yes_no_to_bool , default=a__ , error_message='''Please enter yes or no.''' , ) snake_case_ : Optional[Any] = """Which EC2 instance type you want to use for your training?""" if distributed_type != SageMakerDistributedType.NO: snake_case_ : Optional[int] = _ask_options( a__ , a__ , lambda _UpperCamelCase : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(a__ )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" snake_case_ : List[str] = _ask_field(a__ , lambda _UpperCamelCase : str(a__ ).lower() , default='''ml.p3.2xlarge''' ) snake_case_ : int = 1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): snake_case_ : Dict = _ask_field( '''How many machines do you want use? [1]: ''' , a__ , default=1 , ) snake_case_ : Dict = _ask_options( '''Do you wish to use FP16 or BF16 (mixed precision)?''' , ['''no''', '''fp16''', '''bf16''', '''fp8'''] , _convert_mixed_precision , ) if use_dynamo and mixed_precision == "no": print( '''Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.''' ) return SageMakerConfig( image_uri=a__ , compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER , distributed_type=a__ , use_cpu=a__ , dynamo_config=a__ , eca_instance_type=a__ , profile=a__ , region=a__ , iam_role_name=a__ , mixed_precision=a__ , num_machines=a__ , sagemaker_inputs_file=a__ , sagemaker_metrics_file=a__ , )
279
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def _UpperCAmelCase ( ): '''simple docstring''' a_ : Tuple = ArgumentParser("""Accelerate CLI tool""" , usage="""accelerate <command> [<args>]""" , allow_abbrev=a__) a_ : Any = parser.add_subparsers(help="""accelerate command helpers""") # Register commands get_config_parser(subparsers=a__) env_command_parser(subparsers=a__) launch_command_parser(subparsers=a__) tpu_command_parser(subparsers=a__) test_command_parser(subparsers=a__) # Let's go a_ : Any = parser.parse_args() if not hasattr(a__ , """func"""): parser.print_help() exit(1) # Run args.func(a__) if __name__ == "__main__": main()
248
0
import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _UpperCAmelCase = logging.get_logger('transformers.models.speecht5') def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :Optional[int] , SCREAMING_SNAKE_CASE :Optional[int] , SCREAMING_SNAKE_CASE :List[str] ) -> int: hf_model.apply_weight_norm() __lowerCAmelCase : List[Any] = checkpoint['input_conv.weight_g'] __lowerCAmelCase : Optional[int] = checkpoint['input_conv.weight_v'] __lowerCAmelCase : int = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): __lowerCAmelCase : List[Any] = checkpoint[F'''upsamples.{i}.1.weight_g'''] __lowerCAmelCase : List[str] = checkpoint[F'''upsamples.{i}.1.weight_v'''] __lowerCAmelCase : Any = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): __lowerCAmelCase : Any = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] __lowerCAmelCase : Dict = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] __lowerCAmelCase : Dict = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] __lowerCAmelCase : Optional[int] = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] __lowerCAmelCase : Any = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] __lowerCAmelCase : Optional[int] = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] __lowerCAmelCase : List[Any] = checkpoint['output_conv.1.weight_g'] __lowerCAmelCase : Any = checkpoint['output_conv.1.weight_v'] __lowerCAmelCase : List[Any] = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE :Optional[Any] , SCREAMING_SNAKE_CASE :List[Any] , SCREAMING_SNAKE_CASE :Dict , SCREAMING_SNAKE_CASE :Optional[int]=None , SCREAMING_SNAKE_CASE :Dict=None , ) -> Dict: if config_path is not None: __lowerCAmelCase : Tuple = SpeechTaHifiGanConfig.from_pretrained(_UpperCAmelCase ) else: __lowerCAmelCase : Union[str, Any] = SpeechTaHifiGanConfig() __lowerCAmelCase : Optional[Any] = SpeechTaHifiGan(_UpperCAmelCase ) __lowerCAmelCase : Dict = torch.load(_UpperCAmelCase ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _UpperCAmelCase , _UpperCAmelCase ) __lowerCAmelCase : List[Any] = np.load(_UpperCAmelCase ) __lowerCAmelCase : Optional[int] = stats[0].reshape(-1 ) __lowerCAmelCase : Dict = stats[1].reshape(-1 ) __lowerCAmelCase : str = torch.from_numpy(_UpperCAmelCase ).float() __lowerCAmelCase : str = torch.from_numpy(_UpperCAmelCase ).float() model.save_pretrained(_UpperCAmelCase ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_UpperCAmelCase ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--stats_path', required=True, default=None, type=str, help='Path to stats.npy file') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) _UpperCAmelCase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
350
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = { 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class snake_case_ ( __lowercase ): A_ = 'unispeech-sat' def __init__( self : str , _snake_case : List[Any]=32 , _snake_case : Union[str, Any]=768 , _snake_case : Tuple=12 , _snake_case : Optional[int]=12 , _snake_case : Optional[Any]=3072 , _snake_case : Tuple="gelu" , _snake_case : int=0.1 , _snake_case : List[Any]=0.1 , _snake_case : Union[str, Any]=0.1 , _snake_case : str=0.0 , _snake_case : List[str]=0.0 , _snake_case : int=0.1 , _snake_case : Optional[Any]=0.1 , _snake_case : Optional[Any]=0.02 , _snake_case : int=1E-5 , _snake_case : Dict="group" , _snake_case : Optional[Any]="gelu" , _snake_case : Optional[Any]=(512, 512, 512, 512, 512, 512, 512) , _snake_case : int=(5, 2, 2, 2, 2, 2, 2) , _snake_case : int=(10, 3, 3, 3, 3, 2, 2) , _snake_case : Any=False , _snake_case : Optional[Any]=128 , _snake_case : Tuple=16 , _snake_case : str=False , _snake_case : Dict=True , _snake_case : Tuple=0.05 , _snake_case : str=10 , _snake_case : Tuple=2 , _snake_case : List[Any]=0.0 , _snake_case : str=10 , _snake_case : Any=0 , _snake_case : List[Any]=320 , _snake_case : Union[str, Any]=2 , _snake_case : Dict=0.1 , _snake_case : Dict=100 , _snake_case : Union[str, Any]=256 , _snake_case : int=256 , _snake_case : Union[str, Any]=0.1 , _snake_case : Optional[Any]="mean" , _snake_case : int=False , _snake_case : str=False , _snake_case : str=256 , _snake_case : List[Any]=(512, 512, 512, 512, 1500) , _snake_case : Optional[int]=(5, 3, 3, 1, 1) , _snake_case : Tuple=(1, 2, 3, 1, 1) , _snake_case : Dict=512 , _snake_case : Union[str, Any]=0 , _snake_case : List[str]=1 , _snake_case : Optional[Any]=2 , _snake_case : Optional[int]=504 , **_snake_case : Optional[int] , )->Union[str, Any]: '''simple docstring''' super().__init__(**_snake_case , pad_token_id=_snake_case , bos_token_id=_snake_case , eos_token_id=_snake_case ) __lowerCAmelCase : Dict = hidden_size __lowerCAmelCase : List[Any] = feat_extract_norm __lowerCAmelCase : int = feat_extract_activation __lowerCAmelCase : Union[str, Any] = list(_snake_case ) __lowerCAmelCase : str = list(_snake_case ) __lowerCAmelCase : Optional[Any] = list(_snake_case ) __lowerCAmelCase : Optional[int] = conv_bias __lowerCAmelCase : Dict = num_conv_pos_embeddings __lowerCAmelCase : List[Any] = num_conv_pos_embedding_groups __lowerCAmelCase : Tuple = len(self.conv_dim ) __lowerCAmelCase : int = num_hidden_layers __lowerCAmelCase : str = intermediate_size __lowerCAmelCase : str = hidden_act __lowerCAmelCase : Any = num_attention_heads __lowerCAmelCase : Optional[int] = hidden_dropout __lowerCAmelCase : str = attention_dropout __lowerCAmelCase : int = activation_dropout __lowerCAmelCase : Union[str, Any] = feat_proj_dropout __lowerCAmelCase : List[str] = final_dropout __lowerCAmelCase : Dict = layerdrop __lowerCAmelCase : Tuple = layer_norm_eps __lowerCAmelCase : Optional[Any] = initializer_range __lowerCAmelCase : str = vocab_size __lowerCAmelCase : Optional[int] = num_clusters __lowerCAmelCase : List[Any] = do_stable_layer_norm __lowerCAmelCase : Tuple = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowerCAmelCase : Dict = apply_spec_augment __lowerCAmelCase : List[Any] = mask_time_prob __lowerCAmelCase : List[str] = mask_time_length __lowerCAmelCase : Dict = mask_time_min_masks __lowerCAmelCase : Tuple = mask_feature_prob __lowerCAmelCase : List[str] = mask_feature_length __lowerCAmelCase : str = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __lowerCAmelCase : Optional[int] = num_codevectors_per_group __lowerCAmelCase : List[Any] = num_codevector_groups __lowerCAmelCase : int = contrastive_logits_temperature __lowerCAmelCase : str = feat_quantizer_dropout __lowerCAmelCase : int = num_negatives __lowerCAmelCase : str = codevector_dim __lowerCAmelCase : Any = proj_codevector_dim __lowerCAmelCase : Any = diversity_loss_weight # ctc loss __lowerCAmelCase : Tuple = ctc_loss_reduction __lowerCAmelCase : Any = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. __lowerCAmelCase : Any = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __lowerCAmelCase : List[str] = list(_snake_case ) __lowerCAmelCase : List[str] = list(_snake_case ) __lowerCAmelCase : Optional[int] = list(_snake_case ) __lowerCAmelCase : Optional[int] = xvector_output_dim @property def UpperCAmelCase__ ( self : Optional[Any] )->Any: '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
232
0
from __future__ import annotations def lowercase__ ( __snake_case : int , __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : list[list[int]] = [] create_all_state(1 , __snake_case , __snake_case , [] , __snake_case ) return result def lowercase__ ( __snake_case : int , __snake_case : int , __snake_case : int , __snake_case : list[int] , __snake_case : list[list[int]] , ): '''simple docstring''' if level == 0: total_list.append(current_list[:] ) return for i in range(__snake_case , total_number - level + 2 ): current_list.append(__snake_case ) create_all_state(i + 1 , __snake_case , level - 1 , __snake_case , __snake_case ) current_list.pop() def lowercase__ ( __snake_case : list[list[int]] ): '''simple docstring''' for i in total_list: print(*__snake_case ) if __name__ == "__main__": __UpperCAmelCase = 4 __UpperCAmelCase = 2 __UpperCAmelCase = generate_all_combinations(n, k) print_all_state(total_list)
29
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class __magic_name__ : """simple docstring""" def __init__( self :Tuple , snake_case :Optional[Any] , snake_case :Tuple=13 , snake_case :Dict=7 , snake_case :List[Any]=True , snake_case :List[Any]=True , snake_case :Dict=True , snake_case :Any=True , snake_case :Optional[int]=99 , snake_case :Any=32 , snake_case :Dict=2 , snake_case :int=4 , snake_case :Optional[int]=37 , snake_case :List[str]="gelu" , snake_case :List[Any]=0.1 , snake_case :Optional[Any]=0.1 , snake_case :Tuple=512 , snake_case :Tuple=16 , snake_case :Tuple=2 , snake_case :Optional[int]=0.02 , snake_case :str=3 , snake_case :Optional[int]=4 , snake_case :List[str]=None , snake_case :Tuple=1_000 , ): '''simple docstring''' A_ : str = parent A_ : str = batch_size A_ : str = seq_length A_ : Any = is_training A_ : Any = use_input_mask A_ : str = use_token_type_ids A_ : Tuple = use_labels A_ : Optional[Any] = vocab_size A_ : Dict = hidden_size A_ : str = num_hidden_layers A_ : Dict = num_attention_heads A_ : str = intermediate_size A_ : int = hidden_act A_ : List[Any] = hidden_dropout_prob A_ : Dict = attention_probs_dropout_prob A_ : Optional[Any] = max_position_embeddings A_ : List[Any] = type_vocab_size A_ : Any = type_sequence_label_size A_ : Dict = initializer_range A_ : Any = num_labels A_ : Optional[int] = num_choices A_ : Optional[Any] = scope A_ : Any = range_bbox def SCREAMING_SNAKE_CASE ( self :str ): '''simple docstring''' A_ : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment A_ : Tuple = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: A_ : str = bbox[i, j, 3] A_ : Union[str, Any] = bbox[i, j, 1] A_ : List[Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: A_ : Any = bbox[i, j, 2] A_ : Tuple = bbox[i, j, 0] A_ : int = t A_ : int = tf.convert_to_tensor(snake_case ) A_ : Any = None if self.use_input_mask: A_ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) A_ : str = None if self.use_token_type_ids: A_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A_ : Dict = None A_ : List[Any] = None A_ : List[str] = None if self.use_labels: A_ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A_ : str = ids_tensor([self.batch_size] , self.num_choices ) A_ : int = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self :str , snake_case :Dict , snake_case :Union[str, Any] , snake_case :int , snake_case :int , snake_case :Union[str, Any] , snake_case :Tuple , snake_case :Optional[int] , snake_case :List[Any] ): '''simple docstring''' A_ : Any = TFLayoutLMModel(config=snake_case ) A_ : Tuple = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) A_ : str = model(snake_case , snake_case , token_type_ids=snake_case ) A_ : List[Any] = model(snake_case , snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self :Optional[int] , snake_case :Any , snake_case :List[Any] , snake_case :List[str] , snake_case :Optional[Any] , snake_case :Dict , snake_case :Any , snake_case :Union[str, Any] , snake_case :List[Any] ): '''simple docstring''' A_ : Optional[int] = TFLayoutLMForMaskedLM(config=snake_case ) A_ : Tuple = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self :List[str] , snake_case :Dict , snake_case :Tuple , snake_case :Tuple , snake_case :List[str] , snake_case :Tuple , snake_case :str , snake_case :Optional[int] , snake_case :Any ): '''simple docstring''' A_ : Union[str, Any] = self.num_labels A_ : int = TFLayoutLMForSequenceClassification(config=snake_case ) A_ : Optional[int] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self :Optional[Any] , snake_case :Dict , snake_case :str , snake_case :Optional[Any] , snake_case :int , snake_case :Any , snake_case :Tuple , snake_case :List[str] , snake_case :Union[str, Any] ): '''simple docstring''' A_ : List[Any] = self.num_labels A_ : str = TFLayoutLMForTokenClassification(config=snake_case ) A_ : Union[str, Any] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self :int , snake_case :List[str] , snake_case :Optional[int] , snake_case :Union[str, Any] , snake_case :List[Any] , snake_case :int , snake_case :Any , snake_case :Union[str, Any] , snake_case :Any ): '''simple docstring''' A_ : Optional[Any] = TFLayoutLMForQuestionAnswering(config=snake_case ) A_ : List[Any] = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : int = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ) : Union[str, Any] = config_and_inputs A_ : Optional[Any] = { "input_ids": input_ids, "bbox": bbox, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class __magic_name__ ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" __UpperCamelCase = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) __UpperCamelCase = ( { '''feature-extraction''': TFLayoutLMModel, '''fill-mask''': TFLayoutLMForMaskedLM, '''text-classification''': TFLayoutLMForSequenceClassification, '''token-classification''': TFLayoutLMForTokenClassification, '''zero-shot''': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) __UpperCamelCase = False __UpperCamelCase = True __UpperCamelCase = 10 def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' A_ : Tuple = TFLayoutLMModelTester(self ) A_ : List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def SCREAMING_SNAKE_CASE ( self :Any ): '''simple docstring''' A_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' A_ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) def SCREAMING_SNAKE_CASE ( self :List[Any] ): '''simple docstring''' A_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) @slow def SCREAMING_SNAKE_CASE ( self :Optional[Any] ): '''simple docstring''' for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : List[str] = TFLayoutLMModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip("Onnx compliancy broke with TF 2.10" ) def SCREAMING_SNAKE_CASE ( self :Dict ): '''simple docstring''' pass def __snake_case ( ) -> Optional[Any]: # Here we prepare a batch of 2 sequences to test a LayoutLM forward pass on: # fmt: off A_ : int = tf.convert_to_tensor([[101,1019,1014,1016,1037,12849,4747,1004,14246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,11300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,19274,2772,6205,27814,16147,16147,4343,2047,10283,10969,14389,1012,2338,102]] ) # noqa: E231 A_ : int = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 A_ : Union[str, Any] = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 A_ : List[Any] = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) A_ : Tuple = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class __magic_name__ ( unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE ( self :Tuple ): '''simple docstring''' A_ : str = TFLayoutLMModel.from_pretrained("microsoft/layoutlm-base-uncased" ) A_ , A_ , A_ , A_ , A_ : Tuple = prepare_layoutlm_batch_inputs() # forward pass A_ : Tuple = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the sequence output on [0, :3, :3] A_ : List[Any] = tf.convert_to_tensor( [[0.1785, -0.1947, -0.0425], [-0.3254, -0.2807, 0.2553], [-0.5391, -0.3322, 0.3364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case , atol=1e-3 ) ) # test the pooled output on [1, :3] A_ : Optional[Any] = tf.convert_to_tensor([-0.6580, -0.0214, 0.8552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , snake_case , atol=1e-3 ) ) @slow def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : Union[str, Any] = TFLayoutLMForSequenceClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=2 ) A_ , A_ , A_ , A_ , A_ : Any = prepare_layoutlm_batch_inputs() # forward pass A_ : Dict = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar A_ : List[str] = outputs.loss A_ : Union[str, Any] = (2,) self.assertEqual(loss.shape , snake_case ) # test the shape of the logits A_ : Tuple = outputs.logits A_ : Tuple = (2, 2) self.assertEqual(logits.shape , snake_case ) @slow def SCREAMING_SNAKE_CASE ( self :Optional[int] ): '''simple docstring''' A_ : int = TFLayoutLMForTokenClassification.from_pretrained("microsoft/layoutlm-base-uncased" , num_labels=13 ) A_ , A_ , A_ , A_ , A_ : Optional[int] = prepare_layoutlm_batch_inputs() # forward pass A_ : Union[str, Any] = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) # test the shape of the logits A_ : Dict = outputs.logits A_ : List[Any] = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , snake_case ) @slow def SCREAMING_SNAKE_CASE ( self :List[str] ): '''simple docstring''' A_ : Optional[Any] = TFLayoutLMForQuestionAnswering.from_pretrained("microsoft/layoutlm-base-uncased" ) A_ , A_ , A_ , A_ , A_ : str = prepare_layoutlm_batch_inputs() # forward pass A_ : Union[str, Any] = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the shape of the logits A_ : Union[str, Any] = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , snake_case ) self.assertEqual(outputs.end_logits.shape , snake_case )
300
0
def lowerCamelCase__ ( A : int = 1_00_00_00 ): '''simple docstring''' UpperCAmelCase = 1 UpperCAmelCase = 1 UpperCAmelCase = {1: 1} for inputa in range(2 , A ): UpperCAmelCase = 0 UpperCAmelCase = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: UpperCAmelCase = (3 * number) + 1 counter += 1 if inputa not in counters: UpperCAmelCase = counter if counter > pre_counter: UpperCAmelCase = inputa UpperCAmelCase = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
366
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase : List[str] = { """configuration_time_series_transformer""": [ """TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """TimeSeriesTransformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase : Dict = [ """TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TimeSeriesTransformerForPrediction""", """TimeSeriesTransformerModel""", """TimeSeriesTransformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys _lowercase : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
91
0
from __future__ import annotations from typing import Any def a_ ( __lowercase : list ) -> int: if not postfix_notation: return 0 _snake_case = {'+', '-', '*', '/'} _snake_case = [] for token in postfix_notation: if token in operations: _snake_case , _snake_case = stack.pop(), stack.pop() if token == "+": stack.append(a + b ) elif token == "-": stack.append(a - b ) elif token == "*": stack.append(a * b ) else: if a * b < 0 and a % b != 0: stack.append(a // b + 1 ) else: stack.append(a // b ) else: stack.append(int(__lowercase ) ) return stack.pop() if __name__ == "__main__": import doctest doctest.testmod()
282
import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def A ( self : Optional[int] ): '''simple docstring''' _snake_case = 'hf-internal-testing/tiny-random-t5' _snake_case = AutoTokenizer.from_pretrained(lowercase ) _snake_case = AutoModelForSeqaSeqLM.from_pretrained(lowercase ) _snake_case = tokenizer('This is me' , return_tensors='pt' ) _snake_case = model.to_bettertransformer() self.assertTrue(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) _snake_case = model.generate(**lowercase ) _snake_case = model.reverse_bettertransformer() self.assertFalse(any('BetterTransformer' in mod.__class__.__name__ for _, mod in model.named_modules() ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase ) _snake_case = AutoModelForSeqaSeqLM.from_pretrained(lowercase ) self.assertFalse( any('BetterTransformer' in mod.__class__.__name__ for _, mod in model_reloaded.named_modules() ) ) _snake_case = model_reloaded.generate(**lowercase ) self.assertTrue(torch.allclose(lowercase , lowercase ) ) def A ( self : List[Any] ): '''simple docstring''' _snake_case = 'hf-internal-testing/tiny-random-t5' _snake_case = AutoModelForSeqaSeqLM.from_pretrained(lowercase ) _snake_case = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(lowercase ): model.save_pretrained(lowercase ) _snake_case = model.reverse_bettertransformer() model.save_pretrained(lowercase )
282
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Union[str, Any] = { "configuration_xlm_roberta_xl": [ "XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMRobertaXLConfig", "XLMRobertaXLOnnxConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = [ "XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMRobertaXLForCausalLM", "XLMRobertaXLForMaskedLM", "XLMRobertaXLForMultipleChoice", "XLMRobertaXLForQuestionAnswering", "XLMRobertaXLForSequenceClassification", "XLMRobertaXLForTokenClassification", "XLMRobertaXLModel", "XLMRobertaXLPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaXLConfig, XLMRobertaXLOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta_xl import ( XLM_ROBERTA_XL_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaXLForCausalLM, XLMRobertaXLForMaskedLM, XLMRobertaXLForMultipleChoice, XLMRobertaXLForQuestionAnswering, XLMRobertaXLForSequenceClassification, XLMRobertaXLForTokenClassification, XLMRobertaXLModel, XLMRobertaXLPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
134
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["image_processor", "tokenizer"] __UpperCAmelCase : str = "OwlViTImageProcessor" __UpperCAmelCase : Dict = ("CLIPTokenizer", "CLIPTokenizerFast") def __init__( self : str , lowerCamelCase : Any=None , lowerCamelCase : Any=None , **lowerCamelCase : Union[str, Any] ) -> List[Any]: __snake_case : List[Any] = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowerCamelCase , ) __snake_case : List[Any] = kwargs.pop("feature_extractor" ) __snake_case : str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(lowerCamelCase , lowerCamelCase ) def __call__( self : Union[str, Any] , lowerCamelCase : Tuple=None , lowerCamelCase : int=None , lowerCamelCase : Union[str, Any]=None , lowerCamelCase : List[str]="max_length" , lowerCamelCase : Dict="np" , **lowerCamelCase : str ) -> List[Any]: if text is None and query_images is None and images is None: raise ValueError( "You have to specify at least one text or query image or image. All three cannot be none." ) if text is not None: if isinstance(lowerCamelCase , lowerCamelCase ) or (isinstance(lowerCamelCase , lowerCamelCase ) and not isinstance(text[0] , lowerCamelCase )): __snake_case : Union[str, Any] = [self.tokenizer(lowerCamelCase , padding=lowerCamelCase , return_tensors=lowerCamelCase , **lowerCamelCase )] elif isinstance(lowerCamelCase , lowerCamelCase ) and isinstance(text[0] , lowerCamelCase ): __snake_case : Tuple = [] # Maximum number of queries across batch __snake_case : str = max([len(lowerCamelCase ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(lowerCamelCase ) != max_num_queries: __snake_case : Dict = t + [" "] * (max_num_queries - len(lowerCamelCase )) __snake_case : int = self.tokenizer(lowerCamelCase , padding=lowerCamelCase , return_tensors=lowerCamelCase , **lowerCamelCase ) encodings.append(lowerCamelCase ) else: raise TypeError("Input text should be a string, a list of strings or a nested list of strings" ) if return_tensors == "np": __snake_case : Any = np.concatenate([encoding["input_ids"] for encoding in encodings] , axis=0 ) __snake_case : Tuple = np.concatenate([encoding["attention_mask"] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp __snake_case : List[Any] = jnp.concatenate([encoding["input_ids"] for encoding in encodings] , axis=0 ) __snake_case : Any = jnp.concatenate([encoding["attention_mask"] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch __snake_case : int = torch.cat([encoding["input_ids"] for encoding in encodings] , dim=0 ) __snake_case : int = torch.cat([encoding["attention_mask"] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf __snake_case : int = tf.stack([encoding["input_ids"] for encoding in encodings] , axis=0 ) __snake_case : Dict = tf.stack([encoding["attention_mask"] for encoding in encodings] , axis=0 ) else: raise ValueError("Target return tensor type could not be returned" ) __snake_case : Any = BatchEncoding() __snake_case : Tuple = input_ids __snake_case : int = attention_mask if query_images is not None: __snake_case : List[Any] = BatchEncoding() __snake_case : Union[str, Any] = self.image_processor( lowerCamelCase , return_tensors=lowerCamelCase , **lowerCamelCase ).pixel_values __snake_case : str = query_pixel_values if images is not None: __snake_case : Optional[int] = self.image_processor(lowerCamelCase , return_tensors=lowerCamelCase , **lowerCamelCase ) if text is not None and images is not None: __snake_case : List[str] = image_features.pixel_values return encoding elif query_images is not None and images is not None: __snake_case : int = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**lowerCamelCase ) , tensor_type=lowerCamelCase ) def __snake_case ( self : Dict , *lowerCamelCase : List[Any] , **lowerCamelCase : Union[str, Any] ) -> str: return self.image_processor.post_process(*lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Union[str, Any] , *lowerCamelCase : str , **lowerCamelCase : List[str] ) -> Tuple: return self.image_processor.post_process_object_detection(*lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , *lowerCamelCase : Optional[Any] , **lowerCamelCase : Optional[Any] ) -> Any: return self.image_processor.post_process_image_guided_detection(*lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : List[Any] , *lowerCamelCase : Tuple , **lowerCamelCase : Optional[int] ) -> str: return self.tokenizer.batch_decode(*lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Union[str, Any] , *lowerCamelCase : Tuple , **lowerCamelCase : List[Any] ) -> Tuple: return self.tokenizer.decode(*lowerCamelCase , **lowerCamelCase ) @property def __snake_case ( self : Any ) -> Dict: warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , lowerCamelCase , ) return self.image_processor_class @property def __snake_case ( self : List[str] ) -> Union[str, Any]: warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , lowerCamelCase , ) return self.image_processor
134
1
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): UpperCamelCase_ = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right UpperCamelCase_ = 12_80_22 UpperCamelCase_ = 12_80_28 @require_sentencepiece class a_ (snake_case__ , unittest.TestCase ): __lowerCAmelCase : Optional[int] = MaMaaaTokenizer __lowerCAmelCase : Optional[Any] = False __lowerCAmelCase : str = False __lowerCAmelCase : Any = True def __UpperCamelCase ( self ): super().setUp() _lowerCAmelCase : int = ['</s>', '<unk>', '▁This', '▁is', '▁a', '▁t', 'est', '\u0120', '<pad>'] _lowerCAmelCase : Tuple = dict(zip(_A , range(len(_A ) ) ) ) _lowerCAmelCase : List[Any] = Path(self.tmpdirname ) save_json(_A , save_dir / VOCAB_FILES_NAMES["""vocab_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(_A , save_dir / VOCAB_FILES_NAMES["""spm_file"""] ) _lowerCAmelCase : Optional[int] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __UpperCamelCase ( self , **snake_case_ ): return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **_A ) def __UpperCamelCase ( self , snake_case_ ): return ( "This is a test", "This is a test", ) def __UpperCamelCase ( self ): _lowerCAmelCase : Tuple = '</s>' _lowerCAmelCase : Dict = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_A ) , _A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_A ) , _A ) def __UpperCamelCase ( self ): _lowerCAmelCase : List[Any] = self.get_tokenizer() _lowerCAmelCase : Dict = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """</s>""" ) self.assertEqual(vocab_keys[1] , """<unk>""" ) self.assertEqual(vocab_keys[-1] , """<s>""" ) self.assertEqual(len(_A ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("""Skip this test while all models are still to be uploaded.""" ) def __UpperCamelCase ( self ): pass def __UpperCamelCase ( self ): _lowerCAmelCase : List[str] = self.get_tokenizer() _lowerCAmelCase : List[str] = tokenizer.tokenize("""This is a test""" ) self.assertListEqual(_A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_A ) , [2, 3, 4, 5, 6] , ) _lowerCAmelCase : Optional[int] = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(_A , ["""▁This""", """▁is""", """▁a""", """▁t""", """est"""] ) _lowerCAmelCase : str = tokenizer.convert_tokens_to_string(_A ) self.assertEqual(_A , """This is a test""" ) @slow def __UpperCamelCase ( self ): # fmt: off _lowerCAmelCase : List[Any] = {'input_ids': [[1_2_8_0_2_2, 1_1_0_1_0_8, 3_9_7, 1_1, 3_8_2_7_2, 2_2_4_7, 1_2_4_8_1_1, 2_8_5, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 3_9_5_3_4, 4_4_2_8, 3_9_7, 1_0_1_9, 1_8_1_0_5, 1_5_8_6, 2_0_7, 7, 4_1_3_3_7, 1_6_7_8_6, 2_4_1, 7, 2_0_2_1_4, 1_7, 1_2_5_6_9_0, 1_0_3_9_8, 7, 4_4_3_7_8, 5_8_0_6_9, 6_8_3_4_2, 7_7_9_8, 7_3_4_3, 1_1, 2_9_9, 3_3_3_1_0, 4, 1_5_8, 3_7_3_5_0, 9_4_0_7_7, 4_5_6_9, 2_9_9, 3_3_3_1_0, 9_0, 4, 5_2_8_4_0, 2_9_0, 4, 3_1_2_7_0, 1_1_2, 2_9_9, 6_8_2, 4, 5_2_8_4_0, 3_9_9_5_3, 1_4_0_7_9, 1_9_3, 5_2_5_1_9, 9_0_8_9_4, 1_7_8_9_4, 1_2_0_6_9_7, 1_1, 4_0_4_4_5, 5_5_1, 1_7, 1_0_1_9, 5_2_5_1_9, 9_0_8_9_4, 1_7_7_5_6, 9_6_3, 1_1, 4_0_4_4_5, 4_8_0, 1_7, 9_7_9_2, 1_1_2_0, 5_1_7_3, 1_3_9_3, 6_2_4_0, 1_6_7_8_6, 2_4_1, 1_2_0_9_9_6, 2_8, 1_2_4_5, 1_3_9_3, 1_1_8_2_4_0, 1_1_1_2_3, 1_0_1_9, 9_3_6_1_2, 2_6_9_1, 1_0_6_1_8, 9_8_0_5_8, 1_2_0_4_0_9, 1_9_2_8, 2_7_9, 4, 4_0_6_8_3, 3_6_7, 1_7_8, 2_0_7, 1_0_1_9, 1_0_3, 1_0_3_1_2_1, 5_0_6, 6_5_2_9_6, 5, 2], [1_2_8_0_2_2, 2_1_2_1_7, 3_6_7, 1_1_7, 1_2_5_4_5_0, 1_2_8, 7_1_9, 7, 7_3_0_8, 4_0, 9_3_6_1_2, 1_2_6_6_9, 1_1_1_6, 1_6_7_0_4, 7_1, 1_7_7_8_5, 3_6_9_9, 1_5_5_9_2, 3_5, 1_4_4, 9_5_8_4, 2_4_1, 1_1_9_4_3, 7_1_3, 9_5_0, 7_9_9, 2_2_4_7, 8_8_4_2_7, 1_5_0, 1_4_9, 1_1_8_8_1_3, 1_2_0_7_0_6, 1_0_1_9, 1_0_6_9_0_6, 8_1_5_1_8, 2_8, 1_2_2_4, 2_2_7_9_9, 3_9_7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1_2_8_0_2_2, 1_6_5_8, 1_2_3_3_1_1, 5_1_5_5, 5_5_7_8, 4_7_2_2, 2_7_9, 1_4_9_4_7, 2_3_6_6, 1_1_2_0, 1_1_9_7, 1_4, 1_3_4_8, 9_2_3_2, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_A , model_name="""facebook/m2m100_418M""" , revision="""c168bae485c864188cf9aa0e4108b0b6934dc91e""" , ) @require_torch @require_sentencepiece @require_tokenizers class a_ (unittest.TestCase ): __lowerCAmelCase : List[Any] = '''facebook/m2m100_418M''' __lowerCAmelCase : Any = [ '''In my opinion, there are two levels of response from the French government.''', '''NSA Affair Emphasizes Complete Lack of Debate on Intelligence''', ] __lowerCAmelCase : Optional[Any] = [ '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', ] # fmt: off __lowerCAmelCase : List[str] = [EN_CODE, 5_9_3, 1_9_4_9, 1_1_5_7_8_1, 4, 7_1_5_8_6, 4_2_3_4, 6_0_6_3_3, 1_2_6_2_3_3, 4_3_2, 1_2_3_8_0_8, 1_5_5_9_2, 1_1_9_7, 1_1_7_1_3_2, 1_2_0_6_1_8, 5, 2] @classmethod def __UpperCamelCase ( cls ): _lowerCAmelCase : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="""en""" , tgt_lang="""fr""" ) _lowerCAmelCase : Optional[Any] = 1 return cls def __UpperCamelCase ( self ): self.assertEqual(self.tokenizer.get_lang_id("""ar""" ) , 1_2_8_0_0_6 ) self.assertEqual(self.tokenizer.get_lang_id("""en""" ) , 1_2_8_0_2_2 ) self.assertEqual(self.tokenizer.get_lang_id("""ro""" ) , 1_2_8_0_7_6 ) self.assertEqual(self.tokenizer.get_lang_id("""mr""" ) , 1_2_8_0_6_3 ) def __UpperCamelCase ( self ): _lowerCAmelCase : List[Any] = self.tokenizer.get_vocab() self.assertEqual(len(_A ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["""<unk>"""] , 3 ) self.assertIn(self.tokenizer.get_lang_token("""en""" ) , _A ) def __UpperCamelCase ( self ): _lowerCAmelCase : Any = 'en' _lowerCAmelCase : int = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , _A ) def __UpperCamelCase ( self ): self.assertIn(_A , self.tokenizer.all_special_ids ) # fmt: off _lowerCAmelCase : str = [FR_CODE, 5_3_6_4, 8_2, 8_6_4_2, 4, 2_9_4, 4_7, 8, 1_4_0_2_8, 1_3_6, 3_2_8_6, 9_7_0_6, 6, 9_0_7_9_7, 6, 1_4_4_0_1_2, 1_6_2, 8_8_1_2_8, 3_0_0_6_1, 5, 2] # fmt: on _lowerCAmelCase : str = self.tokenizer.decode(_A , skip_special_tokens=_A ) _lowerCAmelCase : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=_A ) self.assertEqual(_A , _A ) self.assertNotIn(self.tokenizer.eos_token , _A ) def __UpperCamelCase ( self ): _lowerCAmelCase : int = tempfile.mkdtemp() _lowerCAmelCase : int = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(_A ) _lowerCAmelCase : Union[str, Any] = MaMaaaTokenizer.from_pretrained(_A ) self.assertDictEqual(new_tok.lang_token_to_id , _A ) @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : List[Any] = 'en' _lowerCAmelCase : Tuple = 'fr' _lowerCAmelCase : List[str] = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=_A , return_tensors="""pt""" ) _lowerCAmelCase : Union[str, Any] = shift_tokens_right( batch["""labels"""] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: _lowerCAmelCase : str = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : Tuple = 'mr' self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) _lowerCAmelCase : List[Any] = 'zh' self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : int = 'mr' self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""mr""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) _lowerCAmelCase : Optional[Any] = 'zh' self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("""zh""" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def __UpperCamelCase ( self ): _lowerCAmelCase : Any = self.tokenizer._build_translation_inputs("""A test""" , return_tensors="""pt""" , src_lang="""en""" , tgt_lang="""ar""" ) self.assertEqual( nested_simplify(_A ) , { # en_XX, A, test, EOS """input_ids""": [[1_2_8_0_2_2, 5_8, 4_1_8_3, 2]], """attention_mask""": [[1, 1, 1, 1]], # ar_AR """forced_bos_token_id""": 1_2_8_0_0_6, } , )
309
def _SCREAMING_SNAKE_CASE ( a ) -> Tuple: __A , __A : Optional[Any] = [], [] while len(a ) > 1: __A , __A : Any = min(a ), max(a ) start.append(a ) end.append(a ) collection.remove(a ) collection.remove(a ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCAmelCase : int = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase : Dict = [int(item) for item in user_input.split(''',''')] print(*merge_sort(unsorted), sep=''',''')
280
0
'''simple docstring''' import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def a__ ( a__ , a__ , a__ , a__ , a__ = None , a__ = None , a__ = None , ): """simple docstring""" if config_name_or_path is None: __SCREAMING_SNAKE_CASE = """facebook/rag-token-base""" if model_type == """rag_token""" else """facebook/rag-sequence-base""" if generator_tokenizer_name_or_path is None: __SCREAMING_SNAKE_CASE = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: __SCREAMING_SNAKE_CASE = question_encoder_name_or_path __SCREAMING_SNAKE_CASE = RagTokenForGeneration if model_type == """rag_token""" else RagSequenceForGeneration # Save model. __SCREAMING_SNAKE_CASE = RagConfig.from_pretrained(a__ ) __SCREAMING_SNAKE_CASE = AutoConfig.from_pretrained(a__ ) __SCREAMING_SNAKE_CASE = AutoConfig.from_pretrained(a__ ) __SCREAMING_SNAKE_CASE = gen_config __SCREAMING_SNAKE_CASE = question_encoder_config __SCREAMING_SNAKE_CASE = model_class.from_pretrained_question_encoder_generator( a__ , a__ , config=a__ ) rag_model.save_pretrained(a__ ) # Sanity check. model_class.from_pretrained(a__ ) # Save tokenizers. __SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(a__ ) gen_tokenizer.save_pretrained(dest_dir / """generator_tokenizer/""" ) __SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained(a__ ) question_encoder_tokenizer.save_pretrained(dest_dir / """question_encoder_tokenizer/""" ) if __name__ == "__main__": UpperCAmelCase : Tuple = argparse.ArgumentParser() parser.add_argument( '--model_type', choices=['rag_sequence', 'rag_token'], required=True, type=str, help='RAG model type: rag_sequence, rag_token', ) parser.add_argument('--dest', type=str, required=True, help='Path to the output checkpoint directory.') parser.add_argument('--generator_name_or_path', type=str, required=True, help='Generator model identifier') parser.add_argument( '--question_encoder_name_or_path', type=str, required=True, help='Question encoder model identifier' ) parser.add_argument( '--generator_tokenizer_name_or_path', type=str, help='Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``', ) parser.add_argument( '--question_encoder_tokenizer_name_or_path', type=str, help='Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``', ) parser.add_argument( '--config_name_or_path', type=str, help=( 'Identifier of the model config to use, if not provided, resolves to a base config for a given' ' ``model_type``' ), ) UpperCAmelCase : Optional[int] = parser.parse_args() UpperCAmelCase : Dict = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
371
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=13 , __SCREAMING_SNAKE_CASE : Any=7 , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : List[Any]=99 , __SCREAMING_SNAKE_CASE : Union[str, Any]=32 , __SCREAMING_SNAKE_CASE : Dict=5 , __SCREAMING_SNAKE_CASE : str=4 , __SCREAMING_SNAKE_CASE : Tuple=37 , __SCREAMING_SNAKE_CASE : List[Any]="gelu" , __SCREAMING_SNAKE_CASE : Tuple=0.1 , __SCREAMING_SNAKE_CASE : Optional[int]=0.1 , __SCREAMING_SNAKE_CASE : Optional[Any]=512 , __SCREAMING_SNAKE_CASE : Optional[Any]=16 , __SCREAMING_SNAKE_CASE : Optional[Any]=2 , __SCREAMING_SNAKE_CASE : Tuple=0.02 , __SCREAMING_SNAKE_CASE : List[Any]=4 , ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = parent __SCREAMING_SNAKE_CASE = batch_size __SCREAMING_SNAKE_CASE = seq_length __SCREAMING_SNAKE_CASE = is_training __SCREAMING_SNAKE_CASE = use_attention_mask __SCREAMING_SNAKE_CASE = use_token_type_ids __SCREAMING_SNAKE_CASE = use_labels __SCREAMING_SNAKE_CASE = vocab_size __SCREAMING_SNAKE_CASE = hidden_size __SCREAMING_SNAKE_CASE = num_hidden_layers __SCREAMING_SNAKE_CASE = num_attention_heads __SCREAMING_SNAKE_CASE = intermediate_size __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = max_position_embeddings __SCREAMING_SNAKE_CASE = type_vocab_size __SCREAMING_SNAKE_CASE = type_sequence_label_size __SCREAMING_SNAKE_CASE = initializer_range __SCREAMING_SNAKE_CASE = num_choices def UpperCAmelCase__ ( self : Dict ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __SCREAMING_SNAKE_CASE = None if self.use_attention_mask: __SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) __SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: __SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __SCREAMING_SNAKE_CASE = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def UpperCAmelCase__ ( self : List[Any] ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = config_and_inputs __SCREAMING_SNAKE_CASE = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class lowerCAmelCase__ ( a , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ = True lowerCAmelCase__ = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = FlaxRoFormerModelTester(self ) @slow def UpperCAmelCase__ ( self : int ) -> Any: """simple docstring""" for model_class_name in self.all_model_classes: __SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("""junnyu/roformer_chinese_small""" , from_pt=__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) @require_flax class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self : Any ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = FlaxRoFormerForMaskedLM.from_pretrained("""junnyu/roformer_chinese_base""" ) __SCREAMING_SNAKE_CASE = jnp.array([[0, 1, 2, 3, 4, 5]] ) __SCREAMING_SNAKE_CASE = model(__SCREAMING_SNAKE_CASE )[0] __SCREAMING_SNAKE_CASE = 50_000 __SCREAMING_SNAKE_CASE = (1, 6, vocab_size) self.assertEqual(output.shape , __SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = jnp.array( [[[-0.1205, -1.0265, 0.2922], [-1.5134, 0.1974, 0.1519], [-5.0135, -3.9003, -0.8404]]] ) self.assertTrue(jnp.allclose(output[:, :3, :3] , __SCREAMING_SNAKE_CASE , atol=1E-4 ) )
331
0
import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType _lowercase : Optional[List[str]] =None _lowercase : Tuple ="<" if sys.byteorder == "little" else ">" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image _lowercase : Optional[int] =[ np.dtype("|b1"), np.dtype("|u1"), np.dtype("<u2"), np.dtype(">u2"), np.dtype("<i2"), np.dtype(">i2"), np.dtype("<u4"), np.dtype(">u4"), np.dtype("<i4"), np.dtype(">i4"), np.dtype("<f4"), np.dtype(">f4"), np.dtype("<f8"), np.dtype(">f8"), ] @dataclass class snake_case__ : """simple docstring""" __lowerCAmelCase :bool = True __lowerCAmelCase :Optional[str] = None # Automatically constructed __lowerCAmelCase :ClassVar[str] = "PIL.Image.Image" __lowerCAmelCase :ClassVar[Any] = pa.struct({"bytes": pa.binary(), "path": pa.string()} ) __lowerCAmelCase :str = field(default="Image" , init=A__ , repr=A__ ) def __call__( self ) -> str: """simple docstring""" return self.pa_type def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> dict: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""" ) if isinstance(__lowercase , __lowercase ): a__ : Union[str, Any] = np.array(__lowercase ) if isinstance(__lowercase , __lowercase ): return {"path": value, "bytes": None} elif isinstance(__lowercase , __lowercase ): return {"path": None, "bytes": value} elif isinstance(__lowercase , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(__lowercase ) elif isinstance(__lowercase , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(__lowercase ) elif value.get("""path""" ) is not None and os.path.isfile(value["""path"""] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("""path""" )} elif value.get("""bytes""" ) is not None or value.get("""path""" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("""bytes""" ), "path": value.get("""path""" )} else: raise ValueError( F'''An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase=None ) -> "PIL.Image.Image": """simple docstring""" if not self.decode: raise RuntimeError("""Decoding is disabled for this feature. Please use Image(decode=True) instead.""" ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support decoding images, please install 'Pillow'.""" ) if token_per_repo_id is None: a__ : Optional[int] = {} a__ , a__ : Union[str, Any] = value["""path"""], value["""bytes"""] if bytes_ is None: if path is None: raise ValueError(F'''An image should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) else: if is_local_path(__lowercase ): a__ : List[Any] = PIL.Image.open(__lowercase ) else: a__ : Any = path.split("""::""" )[-1] try: a__ : Any = string_to_dict(__lowercase , config.HUB_DATASETS_URL )["""repo_id"""] a__ : str = token_per_repo_id.get(__lowercase ) except ValueError: a__ : Optional[int] = None with xopen(__lowercase , """rb""" , use_auth_token=__lowercase ) as f: a__ : List[Any] = BytesIO(f.read() ) a__ : Any = PIL.Image.open(bytes_ ) else: a__ : Optional[Any] = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def SCREAMING_SNAKE_CASE__( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: """simple docstring""" from .features import Value return ( self if self.decode else { "bytes": Value("""binary""" ), "path": Value("""string""" ), } ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> pa.StructArray: """simple docstring""" if pa.types.is_string(storage.type ): a__ : str = pa.array([None] * len(__lowercase ) , type=pa.binary() ) a__ : str = pa.StructArray.from_arrays([bytes_array, storage] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): a__ : List[str] = pa.array([None] * len(__lowercase ) , type=pa.string() ) a__ : int = pa.StructArray.from_arrays([storage, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("""bytes""" ) >= 0: a__ : List[str] = storage.field("""bytes""" ) else: a__ : Union[str, Any] = pa.array([None] * len(__lowercase ) , type=pa.binary() ) if storage.type.get_field_index("""path""" ) >= 0: a__ : Union[str, Any] = storage.field("""path""" ) else: a__ : List[Any] = pa.array([None] * len(__lowercase ) , type=pa.string() ) a__ : Any = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): a__ : Dict = pa.array( [encode_np_array(np.array(__lowercase ) )["""bytes"""] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) a__ : List[str] = pa.array([None] * len(__lowercase ) , type=pa.string() ) a__ : Dict = pa.StructArray.from_arrays( [bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(__lowercase , self.pa_type ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> pa.StructArray: """simple docstring""" @no_op_if_value_is_null def path_to_bytes(__lowercase ): with xopen(__lowercase , """rb""" ) as f: a__ : Union[str, Any] = f.read() return bytes_ a__ : int = pa.array( [ (path_to_bytes(x["""path"""] ) if x["""bytes"""] is None else x["""bytes"""]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) a__ : Union[str, Any] = pa.array( [os.path.basename(__lowercase ) if path is not None else None for path in storage.field("""path""" ).to_pylist()] , type=pa.string() , ) a__ : Dict = pa.StructArray.from_arrays([bytes_array, path_array] , ["""bytes""", """path"""] , mask=bytes_array.is_null() ) return array_cast(__lowercase , self.pa_type ) def lowerCAmelCase_ ( ) -> List[str]: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""") global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() a__ : int = list(set(PIL.Image.OPEN.keys()) & set(PIL.Image.SAVE.keys())) return _IMAGE_COMPRESSION_FORMATS def lowerCAmelCase_ ( _lowercase : "PIL.Image.Image") -> bytes: """simple docstring""" a__ : int = BytesIO() if image.format in list_image_compression_formats(): a__ : int = image.format else: a__ : List[str] = """PNG""" if image.mode in ["""1""", """L""", """LA""", """RGB""", """RGBA"""] else """TIFF""" image.save(_lowercase , format=_lowercase) return buffer.getvalue() def lowerCAmelCase_ ( _lowercase : "PIL.Image.Image") -> dict: """simple docstring""" if hasattr(_lowercase , """filename""") and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(_lowercase)} def lowerCAmelCase_ ( _lowercase : np.ndarray) -> dict: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""") a__ : str = array.dtype a__ : List[Any] = dtype.byteorder if dtype.byteorder != """=""" else _NATIVE_BYTEORDER a__ : Optional[int] = dtype.kind a__ : Optional[int] = dtype.itemsize a__ : Union[str, Any] = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: a__ : List[Any] = np.dtype("""|u1""") if dtype_kind not in ["u", "i"]: raise TypeError( F'''Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.''') if dtype is not dest_dtype: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''') # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: a__ : Tuple = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: a__ : Optional[int] = dtype_byteorder + dtype_kind + str(_lowercase) a__ : int = np.dtype(_lowercase) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''') break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F'''Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}''') a__ : Optional[int] = PIL.Image.fromarray(array.astype(_lowercase)) return {"path": None, "bytes": image_to_bytes(_lowercase)} def lowerCAmelCase_ ( _lowercase : Union[List[str], List[dict], List[np.ndarray], List["PIL.Image.Image"]]) -> List[dict]: """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("""To support encoding images, please install 'Pillow'.""") if objs: a__ , a__ : Dict = first_non_null_value(_lowercase) if isinstance(_lowercase , _lowercase): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(_lowercase , np.ndarray): a__ : Union[str, Any] = no_op_if_value_is_null(_lowercase) return [obj_to_image_dict_func(_lowercase) for obj in objs] elif isinstance(_lowercase , PIL.Image.Image): a__ : List[str] = no_op_if_value_is_null(_lowercase) return [obj_to_image_dict_func(_lowercase) for obj in objs] else: return objs else: return objs
170
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers import ( TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST, TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST, BertConfig, DPRConfig, TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) class snake_case__ : """simple docstring""" def __init__( self , __lowercase , __lowercase=1_3 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=9_9 , __lowercase=3_2 , __lowercase=2 , __lowercase=4 , __lowercase=3_7 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=5_1_2 , __lowercase=1_6 , __lowercase=2 , __lowercase=0.0_2 , __lowercase=3 , __lowercase=4 , __lowercase=None , __lowercase=0 , ) -> Optional[Any]: """simple docstring""" a__ : Optional[int] = parent a__ : int = batch_size a__ : Dict = seq_length a__ : Optional[Any] = is_training a__ : Optional[Any] = use_input_mask a__ : str = use_token_type_ids a__ : List[Any] = use_labels a__ : int = vocab_size a__ : List[Any] = hidden_size a__ : int = num_hidden_layers a__ : Optional[Any] = num_attention_heads a__ : Tuple = intermediate_size a__ : Dict = hidden_act a__ : Any = hidden_dropout_prob a__ : List[str] = attention_probs_dropout_prob a__ : Optional[Any] = max_position_embeddings a__ : List[Any] = type_vocab_size a__ : Dict = type_sequence_label_size a__ : List[Any] = initializer_range a__ : Dict = num_labels a__ : int = num_choices a__ : Union[str, Any] = scope a__ : str = projection_dim def SCREAMING_SNAKE_CASE__( self ) -> List[str]: """simple docstring""" a__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a__ : Optional[int] = None if self.use_input_mask: # follow test_modeling_tf_ctrl.py a__ : str = random_attention_mask([self.batch_size, self.seq_length] ) a__ : Tuple = None if self.use_token_type_ids: a__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a__ : str = None a__ : List[str] = None a__ : int = None if self.use_labels: a__ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_choices ) a__ : int = BertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) a__ : str = DPRConfig(projection_dim=self.projection_dim , **config.to_dict() ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> str: """simple docstring""" a__ : Tuple = TFDPRContextEncoder(config=__lowercase ) a__ : Optional[int] = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase ) a__ : Dict = model(__lowercase , token_type_ids=__lowercase ) a__ : List[str] = model(__lowercase ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Optional[Any]: """simple docstring""" a__ : List[str] = TFDPRQuestionEncoder(config=__lowercase ) a__ : Optional[Any] = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase ) a__ : str = model(__lowercase , token_type_ids=__lowercase ) a__ : Optional[int] = model(__lowercase ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.projection_dim or self.hidden_size) ) def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase ) -> Union[str, Any]: """simple docstring""" a__ : int = TFDPRReader(config=__lowercase ) a__ : List[Any] = model(__lowercase , attention_mask=__lowercase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.relevance_logits.shape , (self.batch_size,) ) def SCREAMING_SNAKE_CASE__( self ) -> str: """simple docstring""" a__ : Union[str, Any] = self.prepare_config_and_inputs() ( ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ( a__ ) , ) : Union[str, Any] = config_and_inputs a__ : str = {"""input_ids""": input_ids} return config, inputs_dict @require_tf class snake_case__ (A__ , A__ , unittest.TestCase ): """simple docstring""" __lowerCAmelCase :Optional[int] = ( ( TFDPRContextEncoder, TFDPRQuestionEncoder, TFDPRReader, ) if is_tf_available() else () ) __lowerCAmelCase :Dict = {"feature-extraction": TFDPRQuestionEncoder} if is_tf_available() else {} __lowerCAmelCase :List[Any] = False __lowerCAmelCase :Optional[Any] = False __lowerCAmelCase :Dict = False __lowerCAmelCase :int = False __lowerCAmelCase :Optional[Any] = False def SCREAMING_SNAKE_CASE__( self ) -> List[str]: """simple docstring""" a__ : Optional[Any] = TFDPRModelTester(self ) a__ : Dict = ConfigTester(self , config_class=__lowercase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_context_encoder(*__lowercase ) def SCREAMING_SNAKE_CASE__( self ) -> Dict: """simple docstring""" a__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_question_encoder(*__lowercase ) def SCREAMING_SNAKE_CASE__( self ) -> Any: """simple docstring""" a__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_dpr_reader(*__lowercase ) @slow def SCREAMING_SNAKE_CASE__( self ) -> int: """simple docstring""" for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : Optional[Any] = TFDPRContextEncoder.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) for model_name in TF_DPR_CONTEXT_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : str = TFDPRContextEncoder.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) for model_name in TF_DPR_QUESTION_ENCODER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : Optional[Any] = TFDPRQuestionEncoder.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) for model_name in TF_DPR_READER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : int = TFDPRReader.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) @require_tf class snake_case__ (unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE__( self ) -> Tuple: """simple docstring""" a__ : Any = TFDPRQuestionEncoder.from_pretrained("""facebook/dpr-question_encoder-single-nq-base""" ) a__ : Union[str, Any] = tf.constant( [[1_0_1, 7_5_9_2, 1_0_1_0, 2_0_0_3, 2_0_2_6, 3_8_9_9, 1_0_1_4_0, 1_0_2_9, 1_0_2]] ) # [CLS] hello, is my dog cute? [SEP] a__ : Any = model(__lowercase )[0] # embedding shape = (1, 768) # compare the actual values for a slice. a__ : Optional[int] = tf.constant( [ [ 0.0_3_2_3_6_2_5_3, 0.1_2_7_5_3_3_3_5, 0.1_6_8_1_8_5_0_9, 0.0_0_2_7_9_7_8_6, 0.3_8_9_6_9_3_3, 0.2_4_2_6_4_9_4_5, 0.2_1_7_8_9_7_1, -0.0_2_3_3_5_2_2_7, -0.0_8_4_8_1_9_5_9, -0.1_4_3_2_4_1_1_7, ] ] ) self.assertTrue(numpy.allclose(output[:, :1_0].numpy() , expected_slice.numpy() , atol=1E-4 ) )
170
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCamelCase : Any = { """configuration_m2m_100""": ["""M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP""", """M2M100Config""", """M2M100OnnxConfig"""], """tokenization_m2m_100""": ["""M2M100Tokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = [ """M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST""", """M2M100ForConditionalGeneration""", """M2M100Model""", """M2M100PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys __lowerCamelCase : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
140
def A_ ( _lowerCAmelCase = "The quick brown fox jumps over the lazy dog" , ) -> bool: UpperCamelCase : Union[str, Any] = set() # Replace all the whitespace in our sentence UpperCamelCase : Union[str, Any] = input_str.replace(" " , "" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(_lowerCAmelCase ) == 26 def A_ ( _lowerCAmelCase = "The quick brown fox jumps over the lazy dog" , ) -> bool: UpperCamelCase : List[Any] = [False] * 26 for char in input_str: if char.islower(): UpperCamelCase : Tuple = True elif char.isupper(): UpperCamelCase : str = True return all(_lowerCAmelCase ) def A_ ( _lowerCAmelCase = "The quick brown fox jumps over the lazy dog" , ) -> bool: return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def A_ ( ) -> None: from timeit import timeit UpperCamelCase : Tuple = "from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest" print(timeit("is_pangram()" , setup=_lowerCAmelCase ) ) print(timeit("is_pangram_faster()" , setup=_lowerCAmelCase ) ) print(timeit("is_pangram_fastest()" , setup=_lowerCAmelCase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
140
1
import os from pathlib import Path def _lowerCAmelCase ( ) -> Optional[int]: """simple docstring""" from torch.utils.cpp_extension import load snake_case__ : Union[str, Any] = Path(__lowerCAmelCase ).resolve().parent.parent.parent / '''kernels''' / '''deformable_detr''' snake_case__ : List[str] = [ root / filename for filename in [ '''vision.cpp''', os.path.join('''cpu''' , '''ms_deform_attn_cpu.cpp''' ), os.path.join('''cuda''' , '''ms_deform_attn_cuda.cu''' ), ] ] load( '''MultiScaleDeformableAttention''' , __lowerCAmelCase , with_cuda=__lowerCAmelCase , extra_include_paths=[str(__lowerCAmelCase )] , extra_cflags=['''-DWITH_CUDA=1'''] , extra_cuda_cflags=[ '''-DCUDA_HAS_FP16=1''', '''-D__CUDA_NO_HALF_OPERATORS__''', '''-D__CUDA_NO_HALF_CONVERSIONS__''', '''-D__CUDA_NO_HALF2_OPERATORS__''', ] , ) import MultiScaleDeformableAttention as MSDA return MSDA
230
import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets A__ = datasets.logging.get_logger(__name__) A__ = '''\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric", author = "Moosavi, Nafise Sadat and Strube, Michael", booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", month = aug, year = "2016", address = "Berlin, Germany", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/P16-1060", doi = "10.18653/v1/P16-1060", pages = "632--642", } ''' A__ = '''\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. ''' A__ = ''' Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting \'keep_singletons=False\', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs. min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: \'mentions\': mentions \'muc\': MUC metric [Vilain et al, 1995] \'bcub\': B-cubed [Bagga and Baldwin, 1998] \'ceafe\': CEAFe [Luo et al., 2005] \'lea\': LEA [Moosavi and Strube, 2016] \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric(\'coval\') >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\', ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\', ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\', ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\', ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\', ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0} ''' def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=True , __lowerCAmelCase=False , __lowerCAmelCase="dummy_doc" ) -> int: """simple docstring""" snake_case__ : Dict = {doc: key_lines} snake_case__ : Any = {doc: sys_lines} snake_case__ : Dict = {} snake_case__ : List[str] = 0 snake_case__ : Optional[Any] = 0 snake_case__ : Optional[Any] = 0 snake_case__ : Dict = 0 snake_case__ : List[Any] = 0 snake_case__ : List[Any] = 0 snake_case__ , snake_case__ : Tuple = reader.get_doc_mentions(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase ) key_singletons_num += singletons_num if NP_only or min_span: snake_case__ : str = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) snake_case__ , snake_case__ : int = reader.get_doc_mentions(__lowerCAmelCase , sys_doc_lines[doc] , __lowerCAmelCase ) sys_singletons_num += singletons_num if NP_only or min_span: snake_case__ : Union[str, Any] = reader.set_annotated_parse_trees(__lowerCAmelCase , key_doc_lines[doc] , __lowerCAmelCase , __lowerCAmelCase ) if remove_nested: snake_case__ , snake_case__ : Dict = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters snake_case__ , snake_case__ : Optional[int] = reader.remove_nested_coref_mentions(__lowerCAmelCase , __lowerCAmelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters snake_case__ : Any = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : Optional[int] = reader.get_mention_assignments(__lowerCAmelCase , __lowerCAmelCase ) snake_case__ : List[Any] = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( '''Number of removed nested coreferring mentions in the key ''' f"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( '''Number of resulting singleton clusters in the key ''' f"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( f"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ '''files, respectively''' ) return doc_coref_infos def _lowerCAmelCase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: """simple docstring""" snake_case__ : Optional[Any] = get_coref_infos(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) snake_case__ : str = {} snake_case__ : Optional[int] = 0 snake_case__ : List[Any] = 0 for name, metric in metrics: snake_case__ , snake_case__ , snake_case__ : Any = evaluator.evaluate_documents(__lowerCAmelCase , __lowerCAmelCase , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"""{name}/recall""": recall, f"""{name}/precision""": precision, f"""{name}/f1""": fa} ) logger.info( name.ljust(10 ) , f"""Recall: {recall * 100:.2f}""" , f""" Precision: {precision * 100:.2f}""" , f""" F1: {fa * 100:.2f}""" , ) if conll_subparts_num == 3: snake_case__ : int = (conll / 3) * 100 logger.info(f"""CoNLL score: {conll:.2f}""" ) output_scores.update({'''conll_score''': conll} ) return output_scores def _lowerCAmelCase ( __lowerCAmelCase ) -> List[str]: """simple docstring""" snake_case__ : str = False for line in key_lines: if not line.startswith('''#''' ): if len(line.split() ) > 6: snake_case__ : List[Any] = line.split()[5] if not parse_col == "-": snake_case__ : Optional[Any] = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): def __lowerCamelCase ( self :Dict ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''string''' ) ), '''references''': datasets.Sequence(datasets.Value('''string''' ) ), } ) ,codebase_urls=['''https://github.com/ns-moosavi/coval'''] ,reference_urls=[ '''https://github.com/ns-moosavi/coval''', '''https://www.aclweb.org/anthology/P16-1060''', '''http://www.conll.cemantix.org/2012/data.html''', ] ,) def __lowerCamelCase ( self :Any ,__lowercase :List[Any] ,__lowercase :int ,__lowercase :str=True ,__lowercase :Optional[int]=False ,__lowercase :Optional[Any]=False ,__lowercase :Tuple=False ): snake_case__ : Optional[Any] = [ ('''mentions''', evaluator.mentions), ('''muc''', evaluator.muc), ('''bcub''', evaluator.b_cubed), ('''ceafe''', evaluator.ceafe), ('''lea''', evaluator.lea), ] if min_span: snake_case__ : Optional[int] = util.check_gold_parse_annotation(__lowercase ) if not has_gold_parse: raise NotImplementedError('''References should have gold parse annotation to use \'min_span\'.''' ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" snake_case__ : Any = evaluate( key_lines=__lowercase ,sys_lines=__lowercase ,metrics=__lowercase ,NP_only=__lowercase ,remove_nested=__lowercase ,keep_singletons=__lowercase ,min_span=__lowercase ,) return score
230
1
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging UpperCAmelCase_ = logging.get_logger(__name__) class lowercase__ ( __lowerCamelCase ): '''simple docstring''' a : Any = ["audio_values", "audio_mask"] def __init__( self, __magic_name__=2048, __magic_name__=1, __magic_name__=[16, 16], __magic_name__=128, __magic_name__=44100, __magic_name__=86, __magic_name__=2048, __magic_name__=0.0, **__magic_name__, ) -> str: """simple docstring""" super().__init__( feature_size=__magic_name__, sampling_rate=__magic_name__, padding_value=__magic_name__, **__magic_name__, ) UpperCamelCase__ : Optional[int] = spectrogram_length UpperCamelCase__ : Dict = num_channels UpperCamelCase__ : Optional[int] = patch_size UpperCamelCase__ : Optional[int] = feature_size // self.patch_size[1] UpperCamelCase__ : List[Any] = n_fft UpperCamelCase__ : int = sampling_rate // hop_length_to_sampling_rate UpperCamelCase__ : Dict = sampling_rate UpperCamelCase__ : Optional[Any] = padding_value UpperCamelCase__ : Union[str, Any] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2, num_mel_filters=__magic_name__, min_frequency=0.0, max_frequency=2_2050.0, sampling_rate=__magic_name__, norm='''slaney''', mel_scale='''slaney''', ).T def UpperCamelCase__ ( self, __magic_name__ ) -> np.ndarray: """simple docstring""" UpperCamelCase__ : Any = spectrogram( __magic_name__, window_function(self.n_fft, '''hann''' ), frame_length=self.n_fft, hop_length=self.hop_length, power=2.0, mel_filters=self.mel_filters.T, log_mel='''dB''', db_range=80.0, ) UpperCamelCase__ : int = log_spec[:, :-1] UpperCamelCase__ : Optional[Any] = log_spec - 20.0 UpperCamelCase__ : Union[str, Any] = np.clip(log_spec / 40.0, -2.0, 0.0 ) + 1.0 return log_spec def __call__( self, __magic_name__, __magic_name__ = None, __magic_name__ = True, __magic_name__ = None, __magic_name__ = False, __magic_name__ = False, **__magic_name__, ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( '''This feature extractor is set to support sampling rate''' f" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled" f" with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) UpperCamelCase__ : Union[str, Any] = isinstance(__magic_name__, np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"Only mono-channel audio is supported for input to {self}" ) UpperCamelCase__ : Dict = is_batched_numpy or ( isinstance(__magic_name__, (list, tuple) ) and (isinstance(raw_speech[0], (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase__ : str = [np.asarray([speech], dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__magic_name__, np.ndarray ): UpperCamelCase__ : List[str] = np.asarray(__magic_name__, dtype=np.floataa ) elif isinstance(__magic_name__, np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase__ : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase__ : Tuple = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis UpperCamelCase__ : Optional[Any] = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0], __magic_name__ ): UpperCamelCase__ : Optional[int] = [np.asarray(__magic_name__, dtype=np.floataa ) for feature in audio_features] # Create audio attention mask UpperCamelCase__ : str = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: UpperCamelCase__ : List[Any] = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] UpperCamelCase__ : Optional[Any] = np.array(__magic_name__ ).astype(np.floataa ) # convert into correct format for padding UpperCamelCase__ : Any = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch UpperCamelCase__ : int = np.ones([len(__magic_name__ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) UpperCamelCase__ : Optional[int] = padded_audio_features * self.padding_value for i in range(len(__magic_name__ ) ): UpperCamelCase__ : Dict = audio_features[i] UpperCamelCase__ : Optional[Any] = feature # return as BatchFeature if return_attention_mask: UpperCamelCase__ : Tuple = {'''audio_values''': padded_audio_features, '''audio_mask''': audio_mask} else: UpperCamelCase__ : Any = {'''audio_values''': padded_audio_features} UpperCamelCase__ : str = BatchFeature(data=__magic_name__, tensor_type=__magic_name__ ) return encoded_inputs
247
import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class lowercase__ ( __lowerCamelCase ): '''simple docstring''' def __init__( self, __magic_name__=0.01, __magic_name__=1000 ) -> Tuple: """simple docstring""" UpperCamelCase__ : Union[str, Any] = p_stop UpperCamelCase__ : Any = max_length def __iter__( self ) -> int: """simple docstring""" UpperCamelCase__ : Optional[int] = 0 UpperCamelCase__ : Tuple = False while not stop and count < self.max_length: yield count count += 1 UpperCamelCase__ : Union[str, Any] = random.random() < self.p_stop class lowercase__ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self, __magic_name__, __magic_name__, __magic_name__=False, __magic_name__=True ) -> List[Any]: """simple docstring""" UpperCamelCase__ : str = [ BatchSamplerShard(__magic_name__, 2, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) for i in range(2 ) ] UpperCamelCase__ : Union[str, Any] = [list(__magic_name__ ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(__magic_name__ ) for shard in batch_sampler_shards], [len(__magic_name__ ) for e in expected] ) self.assertListEqual(__magic_name__, __magic_name__ ) def UpperCamelCase__ ( self ) -> Optional[int]: """simple docstring""" # Check the shards when the dataset is a round multiple of total batch size. UpperCamelCase__ : Any = BatchSampler(range(24 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) UpperCamelCase__ : List[str] = BatchSampler(range(24 ), batch_size=3, drop_last=__magic_name__ ) # Expected shouldn't change self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. UpperCamelCase__ : Optional[int] = BatchSampler(range(21 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [0, 1, 2]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) UpperCamelCase__ : List[str] = BatchSampler(range(21 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. UpperCamelCase__ : int = BatchSampler(range(22 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Tuple = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 0, 1]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) UpperCamelCase__ : Any = BatchSampler(range(22 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : List[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. UpperCamelCase__ : Union[str, Any] = BatchSampler(range(20 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 0]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [1, 2, 3]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) UpperCamelCase__ : str = BatchSampler(range(20 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : List[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) # Check the shards when the dataset is very small. UpperCamelCase__ : Optional[int] = BatchSampler(range(2 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : int = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) UpperCamelCase__ : Optional[int] = BatchSampler(range(2 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : List[Any] = [[], []] self.check_batch_sampler_shards(__magic_name__, __magic_name__ ) def UpperCamelCase__ ( self ) -> Union[str, Any]: """simple docstring""" # Check the shards when the dataset is a round multiple of batch size. UpperCamelCase__ : List[str] = BatchSampler(range(24 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : str = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) UpperCamelCase__ : List[Any] = BatchSampler(range(24 ), batch_size=4, drop_last=__magic_name__ ) # Expected shouldn't change self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size. UpperCamelCase__ : Tuple = BatchSampler(range(22 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : List[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [0, 1]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) UpperCamelCase__ : Dict = BatchSampler(range(22 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : List[str] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. UpperCamelCase__ : Dict = BatchSampler(range(21 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : str = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 0]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [1, 2]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) UpperCamelCase__ : Tuple = BatchSampler(range(21 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : Tuple = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) # Check the shards when the dataset is very small. UpperCamelCase__ : List[str] = BatchSampler(range(2 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : str = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) UpperCamelCase__ : Optional[Any] = BatchSampler(range(2 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : Tuple = [[], []] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__ ) def UpperCamelCase__ ( self ) -> Tuple: """simple docstring""" # Check the shards when the dataset is a round multiple of total batch size. UpperCamelCase__ : List[Any] = BatchSampler(range(24 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21, 22, 23]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Optional[int] = BatchSampler(range(24 ), batch_size=3, drop_last=__magic_name__ ) # Expected shouldn't change self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. UpperCamelCase__ : Tuple = BatchSampler(range(21 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : List[str] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Optional[Any] = BatchSampler(range(21 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Dict = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. UpperCamelCase__ : Union[str, Any] = BatchSampler(range(22 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Any = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19, 20]], [[3, 4, 5], [9, 10, 11], [15, 16, 17], [21]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : int = BatchSampler(range(22 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : int = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. UpperCamelCase__ : List[Any] = BatchSampler(range(20 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14], [18, 19]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Optional[int] = BatchSampler(range(20 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [12, 13, 14]], [[3, 4, 5], [9, 10, 11], [15, 16, 17]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) # Check the shards when the dataset is very small. UpperCamelCase__ : List[Any] = BatchSampler(range(2 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Dict = [[[0, 1]], []] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Optional[Any] = BatchSampler(range(2 ), batch_size=3, drop_last=__magic_name__ ) UpperCamelCase__ : Tuple = [[], []] self.check_batch_sampler_shards(__magic_name__, __magic_name__, even_batches=__magic_name__ ) def UpperCamelCase__ ( self ) -> str: """simple docstring""" # Check the shards when the dataset is a round multiple of batch size. UpperCamelCase__ : int = BatchSampler(range(24 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : List[str] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19], [22, 23]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Dict = BatchSampler(range(24 ), batch_size=4, drop_last=__magic_name__ ) # Expected shouldn't change self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size. UpperCamelCase__ : List[Any] = BatchSampler(range(22 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : int = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20, 21]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Optional[int] = BatchSampler(range(22 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : Dict = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. UpperCamelCase__ : Any = BatchSampler(range(21 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : Optional[int] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17], [20]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Tuple = BatchSampler(range(21 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : List[Any] = [ [[0, 1], [4, 5], [8, 9], [12, 13], [16, 17]], [[2, 3], [6, 7], [10, 11], [14, 15], [18, 19]], ] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) # Check the shards when the dataset is very small. UpperCamelCase__ : str = BatchSampler(range(2 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : Optional[Any] = [[[0, 1]], []] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) UpperCamelCase__ : Dict = BatchSampler(range(2 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : int = [[], []] self.check_batch_sampler_shards(__magic_name__, __magic_name__, split_batches=__magic_name__, even_batches=__magic_name__ ) def UpperCamelCase__ ( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase__ : Union[str, Any] = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 10, 11], [12, 13]] UpperCamelCase__ : Optional[Any] = [BatchSamplerShard(__magic_name__, 2, __magic_name__, even_batches=__magic_name__ ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ), 3 ) self.assertEqual(len(batch_sampler_shards[1] ), 2 ) self.assertListEqual(list(batch_sampler_shards[0] ), [[0, 1, 2], [5, 6, 7, 8], [12, 13]] ) self.assertListEqual(list(batch_sampler_shards[1] ), [[3, 4], [9, 10, 11]] ) def UpperCamelCase__ ( self, __magic_name__, __magic_name__, __magic_name__, __magic_name__=False, __magic_name__=2, __magic_name__=False ) -> str: """simple docstring""" random.seed(__magic_name__ ) UpperCamelCase__ : Dict = list(__magic_name__ ) UpperCamelCase__ : Optional[int] = [ IterableDatasetShard( __magic_name__, batch_size=__magic_name__, drop_last=__magic_name__, num_processes=__magic_name__, process_index=__magic_name__, split_batches=__magic_name__, ) for i in range(__magic_name__ ) ] UpperCamelCase__ : str = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(__magic_name__ ) iterable_dataset_lists.append(list(__magic_name__ ) ) UpperCamelCase__ : Dict = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size UpperCamelCase__ : Dict = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(__magic_name__ ), len(__magic_name__ ) ) self.assertTrue(len(__magic_name__ ) % shard_batch_size == 0 ) UpperCamelCase__ : int = [] for idx in range(0, len(__magic_name__ ), __magic_name__ ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(__magic_name__ ) < len(__magic_name__ ): reference += reference self.assertListEqual(__magic_name__, reference[: len(__magic_name__ )] ) def UpperCamelCase__ ( self ) -> Dict: """simple docstring""" UpperCamelCase__ : List[Any] = 42 UpperCamelCase__ : Union[str, Any] = RandomIterableDataset() self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) # Edge case with a very small dataset UpperCamelCase__ : Union[str, Any] = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) self.check_iterable_dataset_shards(__magic_name__, __magic_name__, batch_size=4, drop_last=__magic_name__, split_batches=__magic_name__ ) def UpperCamelCase__ ( self ) -> int: """simple docstring""" UpperCamelCase__ : List[str] = BatchSampler(range(16 ), batch_size=4, drop_last=__magic_name__ ) UpperCamelCase__ : List[str] = SkipBatchSampler(__magic_name__, 2 ) self.assertListEqual(list(__magic_name__ ), [[8, 9, 10, 11], [12, 13, 14, 15]] ) def UpperCamelCase__ ( self ) -> List[str]: """simple docstring""" UpperCamelCase__ : List[str] = SkipDataLoader(list(range(16 ) ), batch_size=4, skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader], [[8, 9, 10, 11], [12, 13, 14, 15]] ) def UpperCamelCase__ ( self ) -> int: """simple docstring""" UpperCamelCase__ : Union[str, Any] = DataLoader(list(range(16 ) ), batch_size=4 ) UpperCamelCase__ : Dict = skip_first_batches(__magic_name__, num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader], [[8, 9, 10, 11], [12, 13, 14, 15]] ) def UpperCamelCase__ ( self ) -> Optional[int]: """simple docstring""" UpperCamelCase__ : Any = DataLoaderShard(list(range(16 ) ), batch_size=4 ) for idx, _ in enumerate(__magic_name__ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__magic_name__ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) def UpperCamelCase__ ( self ) -> Dict: """simple docstring""" Accelerator() UpperCamelCase__ : Tuple = DataLoaderDispatcher(range(16 ), batch_size=4 ) for idx, _ in enumerate(__magic_name__ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(__magic_name__ ): self.assertEqual(dataloader.end_of_dataloader, idx == 3 )
247
1
import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _lowercase : Tuple =logging.getLogger(__name__) class snake_case__ (snake_case_ ): """simple docstring""" def __init__( self , __lowercase=-1 ) -> Dict: """simple docstring""" a__ : str = label_idx def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase ) -> List[InputExample]: """simple docstring""" if isinstance(_A , _A ): a__ : int = mode.value a__ : Union[str, Any] = os.path.join(_A , F'''{mode}.txt''' ) a__ : Dict = 1 a__ : Optional[int] = [] with open(_A , encoding="""utf-8""" ) as f: a__ : str = [] a__ : List[str] = [] for line in f: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=F'''{mode}-{guid_index}''' , words=_A , labels=_A ) ) guid_index += 1 a__ : List[str] = [] a__ : List[Any] = [] else: a__ : List[str] = line.split(""" """ ) words.append(splits[0] ) if len(_A ) > 1: labels.append(splits[self.label_idx].replace("""\n""" , """""" ) ) else: # Examples could have no label for mode = "test" labels.append("""O""" ) if words: examples.append(InputExample(guid=F'''{mode}-{guid_index}''' , words=_A , labels=_A ) ) return examples def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase ) -> Dict: """simple docstring""" a__ : int = 0 for line in test_input_reader: if line.startswith("""-DOCSTART-""" ) or line == "" or line == "\n": writer.write(_A ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: a__ : Dict = line.split()[0] + ' ' + preds_list[example_id].pop(0 ) + '\n' writer.write(_A ) else: logger.warning("""Maximum sequence length exceeded: No prediction for \'%s\'.""" , line.split()[0] ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> List[str]: """simple docstring""" if path: with open(_A , """r""" ) as f: a__ : Optional[int] = f.read().splitlines() if "O" not in labels: a__ : Dict = ['O'] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class snake_case__ (snake_case_ ): """simple docstring""" def __init__( self ) -> Tuple: """simple docstring""" super().__init__(label_idx=-2 ) def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> List[str]: """simple docstring""" if path: with open(_A , """r""" ) as f: a__ : Optional[Any] = f.read().splitlines() if "O" not in labels: a__ : Optional[Any] = ['O'] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class snake_case__ (snake_case_ ): """simple docstring""" def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase ) -> List[InputExample]: """simple docstring""" if isinstance(_A , _A ): a__ : Tuple = mode.value a__ : Tuple = os.path.join(_A , F'''{mode}.txt''' ) a__ : Optional[Any] = 1 a__ : List[str] = [] with open(_A , encoding="""utf-8""" ) as f: for sentence in parse_incr(_A ): a__ : int = [] a__ : Dict = [] for token in sentence: words.append(token["""form"""] ) labels.append(token["""upos"""] ) assert len(_A ) == len(_A ) if words: examples.append(InputExample(guid=F'''{mode}-{guid_index}''' , words=_A , labels=_A ) ) guid_index += 1 return examples def SCREAMING_SNAKE_CASE__( self , __lowercase , __lowercase , __lowercase ) -> int: """simple docstring""" a__ : List[str] = 0 for sentence in parse_incr(_A ): a__ : Tuple = preds_list[example_id] a__ : Tuple = '' for token in sentence: out += F'''{token['form']} ({token['upos']}|{s_p.pop(0 )}) ''' out += "\n" writer.write(_A ) example_id += 1 def SCREAMING_SNAKE_CASE__( self , __lowercase ) -> List[str]: """simple docstring""" if path: with open(_A , """r""" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
170
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class SCREAMING_SNAKE_CASE_ ( snake_case_ ): def __init__( self : Union[str, Any] , _A : Any , _A : Dict ) -> Union[str, Any]: """simple docstring""" snake_case_ : str = params snake_case_ : int = np.array(_A ) snake_case_ : Optional[int] = np.array([len(_A ) for t in data] ) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : Tuple , _A : Optional[int] ) -> str: """simple docstring""" return (self.token_ids[index], self.lengths[index]) def __len__( self : List[str] ) -> str: """simple docstring""" return len(self.lengths ) def UpperCAmelCase_ ( self : Dict ) -> str: """simple docstring""" assert len(self.token_ids ) == len(self.lengths ) assert all(self.lengths[i] == len(self.token_ids[i] ) for i in range(len(self.lengths ) ) ) def UpperCAmelCase_ ( self : Any ) -> Optional[Any]: """simple docstring""" snake_case_ : Dict = self.params.max_model_input_size snake_case_ : Tuple = self.lengths > max_len logger.info(F"""Splitting {sum(_A )} too long sequences.""" ) def divide_chunks(_A : Union[str, Any] , _A : Dict ): return [l[i : i + n] for i in range(0 , len(_A ) , _A )] snake_case_ : Dict = [] snake_case_ : Union[str, Any] = [] if self.params.mlm: snake_case_ ,snake_case_ : Optional[int] = self.params.special_tok_ids['cls_token'], self.params.special_tok_ids['sep_token'] else: snake_case_ ,snake_case_ : Any = self.params.special_tok_ids['bos_token'], self.params.special_tok_ids['eos_token'] for seq_, len_ in zip(self.token_ids , self.lengths ): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_ ) new_lengths.append(len_ ) else: snake_case_ : List[Any] = [] for sub_s in divide_chunks(seq_ , max_len - 2 ): if sub_s[0] != cls_id: snake_case_ : Optional[int] = np.insert(_A , 0 , _A ) if sub_s[-1] != sep_id: snake_case_ : Optional[Any] = np.insert(_A , len(_A ) , _A ) assert len(_A ) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(_A ) new_tok_ids.extend(_A ) new_lengths.extend([len(_A ) for l in sub_seqs] ) snake_case_ : Tuple = np.array(_A ) snake_case_ : int = np.array(_A ) def UpperCAmelCase_ ( self : List[str] ) -> List[str]: """simple docstring""" snake_case_ : Tuple = len(self ) snake_case_ : int = self.lengths > 11 snake_case_ : Dict = self.token_ids[indices] snake_case_ : int = self.lengths[indices] snake_case_ : List[Any] = len(self ) logger.info(F"""Remove {init_size - new_size} too short (<=11 tokens) sequences.""" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> int: """simple docstring""" if "unk_token" not in self.params.special_tok_ids: return else: snake_case_ : Optional[Any] = self.params.special_tok_ids['unk_token'] snake_case_ : Dict = len(self ) snake_case_ : str = np.array([np.count_nonzero(a == unk_token_id ) for a in self.token_ids] ) snake_case_ : Any = (unk_occs / self.lengths) < 0.5 snake_case_ : List[Any] = self.token_ids[indices] snake_case_ : int = self.lengths[indices] snake_case_ : Tuple = len(self ) logger.info(F"""Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).""" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: """simple docstring""" if not self.params.is_master: return logger.info(F"""{len(self )} sequences""" ) # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def UpperCAmelCase_ ( self : Optional[int] , _A : Union[str, Any] ) -> List[Any]: """simple docstring""" snake_case_ : Any = [t[0] for t in batch] snake_case_ : int = [t[1] for t in batch] assert len(_A ) == len(_A ) # Max for paddings snake_case_ : str = max(_A ) # Pad token ids if self.params.mlm: snake_case_ : int = self.params.special_tok_ids['pad_token'] else: snake_case_ : Dict = self.params.special_tok_ids['unk_token'] snake_case_ : Dict = [list(t.astype(_A ) ) + [pad_idx] * (max_seq_len_ - len(_A )) for t in token_ids] assert len(tk_ ) == len(_A ) assert all(len(_A ) == max_seq_len_ for t in tk_ ) snake_case_ : Any = torch.tensor(tk_ ) # (bs, max_seq_len_) snake_case_ : Optional[Any] = torch.tensor(_A ) # (bs) return tk_t, lg_t
327
0
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model SCREAMING_SNAKE_CASE : List[Any] = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=None ) -> int: if rng is None: _lowercase : Optional[int] = random.Random() _lowercase : str = 1 for dim in shape: total_dims *= dim _lowercase : int = [] for _ in range(lowerCamelCase_ ): values.append(rng.randint(0 , vocab_size - 1 ) ) _lowercase : List[Any] = np.array(lowerCamelCase_ , dtype=jnp.intaa ).reshape(lowerCamelCase_ ) return output def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_=None ) -> Dict: _lowercase : Tuple = ids_tensor(lowerCamelCase_ , vocab_size=2 , rng=lowerCamelCase_ ) # make sure that at least one token is attended to for each batch _lowercase : int = 1 return attn_mask @require_flax class _lowerCamelCase: lowercase_ : Tuple = None lowercase_ : Dict = () def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase , _lowercase : Dict = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 _lowercase : Optional[Any] = 2 _lowercase : Optional[int] = inputs['input_ids'].shape[-1] // 2 _lowercase : str = inputs['input_ids'][:max_batch_size, :sequence_length] _lowercase : List[str] = jnp.ones_like(lowerCamelCase) _lowercase : List[Any] = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens _lowercase : Optional[int] = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` _lowercase : Tuple = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : str = self._get_input_ids_and_config() _lowercase : List[str] = False _lowercase : Union[str, Any] = max_length _lowercase : List[Any] = 0 for model_class in self.all_generative_model_classes: _lowercase : Any = model_class(lowerCamelCase) _lowercase : List[Any] = model_class.__name__[4:] # Skip the "Flax" at the beginning _lowercase : Any = getattr(lowerCamelCase, lowerCamelCase) _lowercase : List[Any] = pt_model_class(lowerCamelCase).eval() _lowercase : List[str] = load_flax_weights_in_pytorch_model(lowerCamelCase, flax_model.params) _lowercase : Tuple = flax_model.generate(lowerCamelCase).sequences _lowercase : Optional[int] = pt_model.generate(torch.tensor(lowerCamelCase, dtype=torch.long)) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: _lowercase : Optional[int] = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist(), flax_generation_outputs.tolist()) def UpperCamelCase ( self) -> int: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : Optional[Any] = self._get_input_ids_and_config() _lowercase : Optional[int] = False _lowercase : List[Any] = max_length for model_class in self.all_generative_model_classes: _lowercase : List[str] = model_class(lowerCamelCase) _lowercase : Any = model.generate(lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Tuple = jit(model.generate) _lowercase : Any = jit_generate(lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : List[Any] = self._get_input_ids_and_config() _lowercase : Optional[int] = True _lowercase : Tuple = max_length for model_class in self.all_generative_model_classes: _lowercase : str = model_class(lowerCamelCase) _lowercase : int = model.generate(lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Optional[int] = jit(model.generate) _lowercase : Optional[Any] = jit_generate(lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : str = self._get_input_ids_and_config() _lowercase : Optional[Any] = False _lowercase : Tuple = max_length _lowercase : int = 2 for model_class in self.all_generative_model_classes: _lowercase : List[str] = model_class(lowerCamelCase) _lowercase : Dict = model.generate(lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Optional[int] = jit(model.generate) _lowercase : Union[str, Any] = jit_generate(lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : int = self._get_input_ids_and_config() _lowercase : Optional[int] = False _lowercase : Union[str, Any] = max_length _lowercase : Optional[Any] = 2 _lowercase : Dict = 2 for model_class in self.all_generative_model_classes: _lowercase : str = model_class(lowerCamelCase) _lowercase : List[Any] = model.generate(lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[0], input_ids.shape[0] * config.num_return_sequences) def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : List[Any] = self._get_input_ids_and_config() _lowercase : int = True _lowercase : Dict = max_length _lowercase : Optional[int] = 0.8 _lowercase : Union[str, Any] = 10 _lowercase : List[str] = 0.3 _lowercase : Optional[Any] = 1 _lowercase : str = 8 _lowercase : Union[str, Any] = 9 for model_class in self.all_generative_model_classes: _lowercase : Union[str, Any] = model_class(lowerCamelCase) _lowercase : List[Any] = model.generate(lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : int = jit(model.generate) _lowercase : Dict = jit_generate(lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : int = self._get_input_ids_and_config() _lowercase : Optional[Any] = max_length _lowercase : Union[str, Any] = 1 _lowercase : Dict = 8 _lowercase : Optional[Any] = 9 for model_class in self.all_generative_model_classes: _lowercase : Dict = model_class(lowerCamelCase) _lowercase : Optional[Any] = model.generate(lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Any = jit(model.generate) _lowercase : Dict = jit_generate(lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> Optional[int]: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : Optional[int] = self._get_input_ids_and_config() _lowercase : List[Any] = max_length _lowercase : Dict = 2 _lowercase : int = 1 _lowercase : str = 8 _lowercase : Optional[Any] = 9 for model_class in self.all_generative_model_classes: _lowercase : Optional[int] = model_class(lowerCamelCase) _lowercase : Any = model.generate(lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Dict = jit(model.generate) _lowercase : Tuple = jit_generate(lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : Dict = self._get_input_ids_and_config() # pad attention mask on the left _lowercase : List[Any] = attention_mask.at[(0, 0)].set(0) _lowercase : Tuple = False _lowercase : List[str] = max_length for model_class in self.all_generative_model_classes: _lowercase : str = model_class(lowerCamelCase) _lowercase : Dict = model.generate(lowerCamelCase, attention_mask=lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Optional[int] = jit(model.generate) _lowercase : Union[str, Any] = jit_generate(lowerCamelCase, attention_mask=lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : Union[str, Any] = self._get_input_ids_and_config() # pad attention mask on the left _lowercase : Tuple = attention_mask.at[(0, 0)].set(0) _lowercase : int = True _lowercase : List[str] = max_length for model_class in self.all_generative_model_classes: _lowercase : List[str] = model_class(lowerCamelCase) _lowercase : Any = model.generate(lowerCamelCase, attention_mask=lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Any = jit(model.generate) _lowercase : Optional[int] = jit_generate(lowerCamelCase, attention_mask=lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase , _lowercase , _lowercase , _lowercase : Optional[int] = self._get_input_ids_and_config() # pad attention mask on the left _lowercase : Optional[int] = attention_mask.at[(0, 0)].set(0) _lowercase : Optional[int] = 2 _lowercase : Tuple = max_length for model_class in self.all_generative_model_classes: _lowercase : Tuple = model_class(lowerCamelCase) _lowercase : Optional[int] = model.generate(lowerCamelCase, attention_mask=lowerCamelCase).sequences self.assertEqual(generation_outputs.shape[-1], lowerCamelCase) _lowercase : Union[str, Any] = jit(model.generate) _lowercase : List[Any] = jit_generate(lowerCamelCase, attention_mask=lowerCamelCase).sequences self.assertListEqual(generation_outputs.tolist(), jit_generation_outputs.tolist()) @require_flax class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : Any = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-bert') _lowercase : int = FlaxAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-bert-flax-only') _lowercase : str = 'Hello world' _lowercase : List[str] = tokenizer(lowerCamelCase, return_tensors='np').input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(lowerCamelCase, 'do_samples'): model.generate(lowerCamelCase, do_samples=lowerCamelCase) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(lowerCamelCase, 'foo'): _lowercase : str = {'foo': 'bar'} model.generate(lowerCamelCase, **lowerCamelCase)
84
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE : List[Any] = {"configuration_deit": ["DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DeiTConfig", "DeiTOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Dict = ["DeiTFeatureExtractor"] SCREAMING_SNAKE_CASE : Dict = ["DeiTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : str = [ "DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "DeiTForImageClassification", "DeiTForImageClassificationWithTeacher", "DeiTForMaskedImageModeling", "DeiTModel", "DeiTPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[str] = [ "TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDeiTForImageClassification", "TFDeiTForImageClassificationWithTeacher", "TFDeiTForMaskedImageModeling", "TFDeiTModel", "TFDeiTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_deit import DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, DeiTConfig, DeiTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_deit import DeiTFeatureExtractor from .image_processing_deit import DeiTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deit import ( DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, DeiTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deit import ( TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, TFDeiTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
84
1
"""simple docstring""" import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer _A : Dict = """bart""" _A : List[Any] = True @st.cache(allow_output_mutation=__snake_case ) def __magic_name__ ( ) -> str: if LOAD_DENSE_INDEX: lowercase : Any = AutoTokenizer.from_pretrained("yjernite/retribert-base-uncased" ) lowercase : Union[str, Any] = AutoModel.from_pretrained("yjernite/retribert-base-uncased" ).to("cuda:0" ) lowercase : Optional[int] = qar_model.eval() else: lowercase , lowercase : int = (None, None) if MODEL_TYPE == "bart": lowercase : List[str] = AutoTokenizer.from_pretrained("yjernite/bart_eli5" ) lowercase : Any = AutoModelForSeqaSeqLM.from_pretrained("yjernite/bart_eli5" ).to("cuda:0" ) lowercase : List[str] = torch.load("seq2seq_models/eli5_bart_model_blm_2.pth" ) sas_model.load_state_dict(save_dict["model"] ) lowercase : str = sas_model.eval() else: lowercase , lowercase : Tuple = make_qa_sas_model( model_name="t5-small" , from_file="seq2seq_models/eli5_t5_model_1024_4.pth" , device="cuda:0" ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=__snake_case ) def __magic_name__ ( ) -> Optional[int]: if LOAD_DENSE_INDEX: lowercase : Tuple = faiss.StandardGpuResources() lowercase : str = datasets.load_dataset(path="wiki_snippets" , name="wiki40b_en_100_0" )["train"] lowercase : Any = np.memmap( "wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat" , dtype="float32" , mode="r" , shape=(wikiaab_passages.num_rows, 128) , ) lowercase : Union[str, Any] = faiss.IndexFlatIP(128 ) lowercase : int = faiss.index_cpu_to_gpu(__snake_case , 1 , __snake_case ) wikiaab_gpu_index_flat.add(__snake_case ) # TODO fix for larger GPU else: lowercase , lowercase : List[Any] = (None, None) lowercase : List[str] = Elasticsearch([{"host": "localhost", "port": "9200"}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=__snake_case ) def __magic_name__ ( ) -> int: lowercase : List[Any] = datasets.load_dataset("eli5" , name="LFQA_reddit" ) lowercase : Union[str, Any] = elia["train_eli5"] lowercase : Union[str, Any] = np.memmap( "eli5_questions_reps.dat" , dtype="float32" , mode="r" , shape=(elia_train.num_rows, 128) ) lowercase : Any = faiss.IndexFlatIP(128 ) eli5_train_q_index.add(__snake_case ) return (elia_train, eli5_train_q_index) _A , _A , _A : int = load_indexes() _A , _A , _A , _A : Optional[Any] = load_models() _A , _A : int = load_train_data() def __magic_name__ ( __snake_case : List[str] , __snake_case : Any=10 ) -> Dict: lowercase : List[Any] = embed_questions_for_retrieval([question] , __snake_case , __snake_case ) lowercase , lowercase : str = eli5_train_q_index.search(__snake_case , __snake_case ) lowercase : Union[str, Any] = [elia_train[int(__snake_case )] for i in I[0]] return nn_examples def __magic_name__ ( __snake_case : int , __snake_case : Any="wiki40b" , __snake_case : str="dense" , __snake_case : Tuple=10 ) -> Tuple: if source == "none": lowercase , lowercase : int = (" <P> ".join(["" for _ in range(11 )] ).strip(), []) else: if method == "dense": lowercase , lowercase : Optional[Any] = query_qa_dense_index( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) else: lowercase , lowercase : Optional[int] = query_es_index( __snake_case , __snake_case , index_name="english_wiki40b_snippets_100w" , n_results=__snake_case , ) lowercase : Union[str, Any] = [ (res["article_title"], res["section_title"].strip(), res["score"], res["passage_text"]) for res in hit_lst ] lowercase : List[str] = "question: {} context: {}".format(__snake_case , __snake_case ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda __snake_case : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda __snake_case : None), } ) def __magic_name__ ( __snake_case : int , __snake_case : Tuple , __snake_case : Tuple , __snake_case : Union[str, Any]=64 , __snake_case : Any=256 , __snake_case : Any=False , __snake_case : str=2 , __snake_case : Optional[int]=0.95 , __snake_case : List[str]=0.8 ) -> Optional[Any]: with torch.no_grad(): lowercase : Dict = qa_sas_generate( __snake_case , __snake_case , __snake_case , num_answers=1 , num_beams=__snake_case , min_len=__snake_case , max_len=__snake_case , do_sample=__snake_case , temp=__snake_case , top_p=__snake_case , top_k=__snake_case , max_input_length=1024 , device="cuda:0" , )[0] return (answer, support_list) st.title("""Long Form Question Answering with ELI5""") # Start sidebar _A : Any = """<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" _A : Optional[Any] = """ <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia _A : List[Any] = """ This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) _A : int = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] _A : int = st.sidebar.checkbox("""Demo options""") if demo_options: _A : Tuple = st.sidebar.selectbox( """""", action_list, index=3, ) _A : Optional[int] = action_list.index(action_st) _A : Optional[int] = st.sidebar.selectbox( """""", ["""Show full text of passages""", """Show passage section titles"""], index=0, ) _A : Optional[int] = show_type == """Show full text of passages""" else: _A : Optional[int] = 3 _A : int = True _A : Tuple = st.sidebar.checkbox("""Retrieval options""") if retrieval_options: _A : int = """ ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) _A : List[Any] = st.sidebar.selectbox("""Which Wikipedia format should the model use?""", ["""wiki40b""", """none"""]) _A : List[str] = st.sidebar.selectbox("""Which Wikipedia indexer should the model use?""", ["""dense""", """sparse""", """mixed"""]) else: _A : List[Any] = """wiki40b""" _A : Dict = """dense""" _A : Any = """beam""" _A : Union[str, Any] = 2 _A : Optional[Any] = 64 _A : Union[str, Any] = 2_56 _A : Optional[Any] = None _A : Optional[Any] = None _A : List[Any] = st.sidebar.checkbox("""Generation options""") if generate_options: _A : Any = """ ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) _A : Dict = st.sidebar.selectbox("""Would you like to use beam search or sample an answer?""", ["""beam""", """sampled"""]) _A : Dict = st.sidebar.slider( """Minimum generation length""", min_value=8, max_value=2_56, value=64, step=8, format=None, key=None ) _A : Tuple = st.sidebar.slider( """Maximum generation length""", min_value=64, max_value=5_12, value=2_56, step=16, format=None, key=None ) if sampled == "beam": _A : List[Any] = st.sidebar.slider("""Beam size""", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: _A : Any = st.sidebar.slider( """Nucleus sampling p""", min_value=0.1, max_value=1.0, value=0.9_5, step=0.0_1, format=None, key=None ) _A : Any = st.sidebar.slider( """Temperature""", min_value=0.1, max_value=1.0, value=0.7, step=0.0_1, format=None, key=None ) _A : List[Any] = None # start main text _A : Union[str, Any] = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] _A : int = st.selectbox( """What would you like to ask? ---- select <MY QUESTION> to enter a new query""", questions_list, index=1, ) if question_s == "<MY QUESTION>": _A : Dict = st.text_input("""Enter your question here:""", """""") else: _A : Optional[int] = question_s if st.button("""Show me!"""): if action in [0, 1, 3]: if index_type == "mixed": _A , _A : str = make_support(question, source=wiki_source, method="""dense""", n_results=10) _A , _A : Union[str, Any] = make_support(question, source=wiki_source, method="""sparse""", n_results=10) _A : Any = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] _A : Union[str, Any] = support_list[:10] _A : Optional[Any] = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: _A , _A : Optional[Any] = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: _A , _A : Optional[Any] = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == """sampled"""), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("""### The model generated answer is:""") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("""--- \n ### The model is drawing information from the following Wikipedia passages:""") for i, res in enumerate(support_list): _A : List[str] = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(""" """, """_""")) _A : List[str] = res[1].strip() if sec_titles == "": _A : List[Any] = """[{}]({})""".format(res[0], wiki_url) else: _A : List[Any] = sec_titles.split(""" & """) _A : str = """ & """.join( ["""[{}]({}#{})""".format(sec.strip(), wiki_url, sec.strip().replace(""" """, """_""")) for sec in sec_list] ) st.markdown( """{0:02d} - **Article**: {1:<18} <br> _Section_: {2}""".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( """> <span style=\"font-family:arial; font-size:10pt;\">""" + res[-1] + """</span>""", unsafe_allow_html=True ) if action in [2, 3]: _A : Union[str, Any] = find_nearest_training(question) _A : List[Any] = nn_train_list[0] st.markdown( """--- \n ### The most similar question in the ELI5 training set was: \n\n {}""".format(train_exple["""title"""]) ) _A : Optional[int] = [ """{}. {}""".format(i + 1, """ \n""".join([line.strip() for line in ans.split("""\n""") if line.strip() != """"""])) for i, (ans, sc) in enumerate(zip(train_exple["""answers"""]["""text"""], train_exple["""answers"""]["""score"""])) if i == 0 or sc > 2 ] st.markdown("""##### Its answers were: \n\n {}""".format("""\n""".join(answers_st))) _A : List[str] = """ --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
202
"""simple docstring""" from __future__ import annotations def __magic_name__ ( __snake_case : list[int] ) -> list[int]: if len(__snake_case ) == 0: return array lowercase , lowercase : Tuple = min(__snake_case ), max(__snake_case ) # Compute the variables lowercase : Optional[Any] = _max - _min + 1 lowercase , lowercase : List[str] = [0] * holes_range, [0] * holes_range # Make the sorting. for i in array: lowercase : Tuple = i - _min lowercase : str = i holes_repeat[index] += 1 # Makes the array back by replacing the numbers. lowercase : Union[str, Any] = 0 for i in range(__snake_case ): while holes_repeat[i] > 0: lowercase : Tuple = holes[i] index += 1 holes_repeat[i] -= 1 # Returns the sorted array. return array if __name__ == "__main__": import doctest doctest.testmod() _A : str = input("""Enter numbers separated by comma:\n""") _A : Optional[Any] = [int(x) for x in user_input.split(""",""")] print(pigeon_sort(unsorted))
202
1
"""simple docstring""" import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def lowercase__ ( _UpperCAmelCase ) -> List[str]: '''simple docstring''' lowercase : int = [] embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight''', f'''stage{idx}.patch_embed.proj.weight''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias''', f'''stage{idx}.patch_embed.proj.bias''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight''', f'''stage{idx}.patch_embed.norm.weight''', ) ) embed.append( ( f'''cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias''', f'''stage{idx}.patch_embed.norm.bias''', ) ) return embed def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: '''simple docstring''' lowercase : Optional[Any] = [] attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked''', f'''stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_q.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_q.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_k.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_k.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj_v.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj_v.bias''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight''', f'''stage{idx}.blocks.{cnt}.attn.proj.weight''', ) ) attention_weights.append( ( f'''cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias''', f'''stage{idx}.blocks.{cnt}.attn.proj.bias''', ) ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight''', f'''stage{idx}.blocks.{cnt}.mlp.fc1.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias''', f'''stage{idx}.blocks.{cnt}.mlp.fc1.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight''', f'''stage{idx}.blocks.{cnt}.mlp.fc2.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias''', f'''stage{idx}.blocks.{cnt}.mlp.fc2.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight''', f'''stage{idx}.blocks.{cnt}.norm1.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias''', f'''stage{idx}.blocks.{cnt}.norm1.bias''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight''', f'''stage{idx}.blocks.{cnt}.norm2.weight''') ) attention_weights.append( (f'''cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias''', f'''stage{idx}.blocks.{cnt}.norm2.bias''') ) return attention_weights def lowercase__ ( _UpperCAmelCase ) -> Optional[int]: '''simple docstring''' lowercase : Optional[int] = [] token.append((f'''cvt.encoder.stages.{idx}.cls_token''', 'stage2.cls_token') ) return token def lowercase__ ( ) -> List[Any]: '''simple docstring''' lowercase : int = [] head.append(('layernorm.weight', 'norm.weight') ) head.append(('layernorm.bias', 'norm.bias') ) head.append(('classifier.weight', 'head.weight') ) head.append(('classifier.bias', 'head.bias') ) return head def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Any: '''simple docstring''' lowercase : Dict = 'imagenet-1k-id2label.json' lowercase : int = 10_00 lowercase : int = 'huggingface/label-files' lowercase : Optional[int] = num_labels lowercase : Union[str, Any] = json.load(open(cached_download(hf_hub_url(_UpperCAmelCase , _UpperCAmelCase , repo_type='dataset' ) ) , 'r' ) ) lowercase : List[str] = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} lowercase : Union[str, Any] = idalabel lowercase : Optional[Any] = {v: k for k, v in idalabel.items()} lowercase : Tuple = CvtConfig(num_labels=_UpperCAmelCase , idalabel=_UpperCAmelCase , labelaid=_UpperCAmelCase ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit('/' , 1 )[-1][4:6] == "13": lowercase : str = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit('/' , 1 )[-1][4:6] == "21": lowercase : List[str] = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: lowercase : str = [2, 2, 20] lowercase : Optional[int] = [3, 12, 16] lowercase : List[Any] = [1_92, 7_68, 10_24] lowercase : Union[str, Any] = CvtForImageClassification(_UpperCAmelCase ) lowercase : Any = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' ) lowercase : int = image_size lowercase : Optional[Any] = torch.load(_UpperCAmelCase , map_location=torch.device('cpu' ) ) lowercase : List[str] = OrderedDict() lowercase : Dict = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: lowercase : Tuple = list_of_state_dict + cls_token(_UpperCAmelCase ) lowercase : Union[str, Any] = list_of_state_dict + embeddings(_UpperCAmelCase ) for cnt in range(config.depth[idx] ): lowercase : int = list_of_state_dict + attention(_UpperCAmelCase , _UpperCAmelCase ) lowercase : List[Any] = list_of_state_dict + final() for gg in list_of_state_dict: print(_UpperCAmelCase ) for i in range(len(_UpperCAmelCase ) ): lowercase : Union[str, Any] = original_weights[list_of_state_dict[i][1]] model.load_state_dict(_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) image_processor.save_pretrained(_UpperCAmelCase ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": _UpperCamelCase: int = argparse.ArgumentParser() parser.add_argument( '--cvt_model', default='cvt-w24', type=str, help='Name of the cvt model you\'d like to convert.', ) parser.add_argument( '--image_size', default=3_8_4, type=int, help='Input Image Size', ) parser.add_argument( '--cvt_file_name', default=R'cvtmodels\CvT-w24-384x384-IN-22k.pth', type=str, help='Input Image Size', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _UpperCamelCase: Optional[Any] = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
369
"""simple docstring""" _UpperCamelCase: Dict = 2_5_6 # Modulus to hash a string _UpperCamelCase: Union[str, Any] = 1_0_0_0_0_0_3 def lowercase__ ( _UpperCAmelCase , _UpperCAmelCase ) -> bool: '''simple docstring''' lowercase : Dict = len(_UpperCAmelCase ) lowercase : Union[str, Any] = len(_UpperCAmelCase ) if p_len > t_len: return False lowercase : Union[str, Any] = 0 lowercase : Dict = 0 lowercase : Any = 1 # Calculating the hash of pattern and substring of text for i in range(_UpperCAmelCase ): lowercase : Dict = (ord(pattern[i] ) + p_hash * alphabet_size) % modulus lowercase : Tuple = (ord(text[i] ) + text_hash * alphabet_size) % modulus if i == p_len - 1: continue lowercase : Tuple = (modulus_power * alphabet_size) % modulus for i in range(0 , t_len - p_len + 1 ): if text_hash == p_hash and text[i : i + p_len] == pattern: return True if i == t_len - p_len: continue # Calculate the https://en.wikipedia.org/wiki/Rolling_hash lowercase : str = ( (text_hash - ord(text[i] ) * modulus_power) * alphabet_size + ord(text[i + p_len] ) ) % modulus return False def lowercase__ ( ) -> None: '''simple docstring''' lowercase : Any = 'abc1abc12' lowercase : int = 'alskfjaldsabc1abc1abc12k23adsfabcabc' lowercase : Optional[int] = 'alskfjaldsk23adsfabcabc' assert rabin_karp(_UpperCAmelCase , _UpperCAmelCase ) and not rabin_karp(_UpperCAmelCase , _UpperCAmelCase ) # Test 2) lowercase : str = 'ABABX' lowercase : Tuple = 'ABABZABABYABABX' assert rabin_karp(_UpperCAmelCase , _UpperCAmelCase ) # Test 3) lowercase : int = 'AAAB' lowercase : Union[str, Any] = 'ABAAAAAB' assert rabin_karp(_UpperCAmelCase , _UpperCAmelCase ) # Test 4) lowercase : Union[str, Any] = 'abcdabcy' lowercase : List[str] = 'abcxabcdabxabcdabcdabcy' assert rabin_karp(_UpperCAmelCase , _UpperCAmelCase ) # Test 5) lowercase : Dict = 'Lü' lowercase : Dict = 'Lüsai' assert rabin_karp(_UpperCAmelCase , _UpperCAmelCase ) lowercase : List[Any] = 'Lue' assert not rabin_karp(_UpperCAmelCase , _UpperCAmelCase ) print('Success.' ) if __name__ == "__main__": test_rabin_karp()
53
0
"""simple docstring""" import shutil import tempfile import unittest from unittest.mock import patch from transformers import ( DefaultFlowCallback, IntervalStrategy, PrinterCallback, ProgressCallback, Trainer, TrainerCallback, TrainingArguments, is_torch_available, ) from transformers.testing_utils import require_torch if is_torch_available(): from transformers.trainer import DEFAULT_CALLBACKS from .test_trainer import RegressionDataset, RegressionModelConfig, RegressionPreTrainedModel class SCREAMING_SNAKE_CASE_ ( __a ): """simple docstring""" def __init__( self): __SCREAMING_SNAKE_CASE = [] def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_init_end""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_train_begin""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_train_end""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_epoch_begin""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_epoch_end""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_step_begin""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_step_end""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_evaluate""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_predict""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_save""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_log""") def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__): self.events.append("""on_prediction_step""") @require_torch class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" def snake_case_ ( self): __SCREAMING_SNAKE_CASE = tempfile.mkdtemp() def snake_case_ ( self): shutil.rmtree(self.output_dir) def snake_case_ ( self , lowerCAmelCase__=0 , lowerCAmelCase__=0 , lowerCAmelCase__=6_4 , lowerCAmelCase__=6_4 , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__): # disable_tqdm in TrainingArguments has a flaky default since it depends on the level of logging. We make sure # its set to False since the tests later on depend on its value. __SCREAMING_SNAKE_CASE = RegressionDataset(length=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = RegressionDataset(length=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = RegressionModelConfig(a=lowerCAmelCase__ , b=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = RegressionPreTrainedModel(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = TrainingArguments(self.output_dir , disable_tqdm=lowerCAmelCase__ , report_to=[] , **lowerCAmelCase__) return Trainer( lowerCAmelCase__ , lowerCAmelCase__ , train_dataset=lowerCAmelCase__ , eval_dataset=lowerCAmelCase__ , callbacks=lowerCAmelCase__ , ) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__): self.assertEqual(len(lowerCAmelCase__) , len(lowerCAmelCase__)) # Order doesn't matter __SCREAMING_SNAKE_CASE = sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__: cb.__name__ if isinstance(lowerCAmelCase__ , lowerCAmelCase__) else cb.__class__.__name__) __SCREAMING_SNAKE_CASE = sorted(lowerCAmelCase__ , key=lambda lowerCAmelCase__: cb.__name__ if isinstance(lowerCAmelCase__ , lowerCAmelCase__) else cb.__class__.__name__) for cba, cba in zip(lowerCAmelCase__ , lowerCAmelCase__): if isinstance(lowerCAmelCase__ , lowerCAmelCase__) and isinstance(lowerCAmelCase__ , lowerCAmelCase__): self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__) and not isinstance(lowerCAmelCase__ , lowerCAmelCase__): self.assertEqual(lowerCAmelCase__ , cba.__class__) elif not isinstance(lowerCAmelCase__ , lowerCAmelCase__) and isinstance(lowerCAmelCase__ , lowerCAmelCase__): self.assertEqual(cba.__class__ , lowerCAmelCase__) else: self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__) def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = ["""on_init_end""", """on_train_begin"""] __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = len(trainer.get_eval_dataloader()) __SCREAMING_SNAKE_CASE = ["""on_prediction_step"""] * len(trainer.get_eval_dataloader()) + ["""on_log""", """on_evaluate"""] for _ in range(trainer.state.num_train_epochs): expected_events.append("""on_epoch_begin""") for _ in range(lowerCAmelCase__): step += 1 expected_events += ["on_step_begin", "on_step_end"] if step % trainer.args.logging_steps == 0: expected_events.append("""on_log""") if trainer.args.evaluation_strategy == IntervalStrategy.STEPS and step % trainer.args.eval_steps == 0: expected_events += evaluation_events.copy() if step % trainer.args.save_steps == 0: expected_events.append("""on_save""") expected_events.append("""on_epoch_end""") if trainer.args.evaluation_strategy == IntervalStrategy.EPOCH: expected_events += evaluation_events.copy() expected_events += ["on_log", "on_train_end"] return expected_events def snake_case_ ( self): __SCREAMING_SNAKE_CASE = self.get_trainer() __SCREAMING_SNAKE_CASE = DEFAULT_CALLBACKS.copy() + [ProgressCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) # Callbacks passed at init are added to the default callbacks __SCREAMING_SNAKE_CASE = self.get_trainer(callbacks=[MyTestTrainerCallback]) expected_callbacks.append(lowerCAmelCase__) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) # TrainingArguments.disable_tqdm controls if use ProgressCallback or PrinterCallback __SCREAMING_SNAKE_CASE = self.get_trainer(disable_tqdm=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = DEFAULT_CALLBACKS.copy() + [PrinterCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = DEFAULT_CALLBACKS.copy() + [ProgressCallback] __SCREAMING_SNAKE_CASE = self.get_trainer() # We can add, pop, or remove by class name trainer.remove_callback(lowerCAmelCase__) expected_callbacks.remove(lowerCAmelCase__) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self.get_trainer() __SCREAMING_SNAKE_CASE = trainer.pop_callback(lowerCAmelCase__) self.assertEqual(cb.__class__ , lowerCAmelCase__) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) trainer.add_callback(lowerCAmelCase__) expected_callbacks.insert(0 , lowerCAmelCase__) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) # We can also add, pop, or remove by instance __SCREAMING_SNAKE_CASE = self.get_trainer() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[0] trainer.remove_callback(lowerCAmelCase__) expected_callbacks.remove(lowerCAmelCase__) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self.get_trainer() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[0] __SCREAMING_SNAKE_CASE = trainer.pop_callback(lowerCAmelCase__) self.assertEqual(lowerCAmelCase__ , lowerCAmelCase__) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) trainer.add_callback(lowerCAmelCase__) expected_callbacks.insert(0 , lowerCAmelCase__) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowerCAmelCase__) def snake_case_ ( self): import warnings # XXX: for now ignore scatter_gather warnings in this test since it's not relevant to what's being tested warnings.simplefilter(action="""ignore""" , category=lowerCAmelCase__) __SCREAMING_SNAKE_CASE = self.get_trainer(callbacks=[MyTestTrainerCallback]) trainer.train() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowerCAmelCase__ , self.get_expected_events(lowerCAmelCase__)) # Independent log/save/eval __SCREAMING_SNAKE_CASE = self.get_trainer(callbacks=[MyTestTrainerCallback] , logging_steps=5) trainer.train() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowerCAmelCase__ , self.get_expected_events(lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = self.get_trainer(callbacks=[MyTestTrainerCallback] , save_steps=5) trainer.train() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowerCAmelCase__ , self.get_expected_events(lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = self.get_trainer(callbacks=[MyTestTrainerCallback] , eval_steps=5 , evaluation_strategy="""steps""") trainer.train() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowerCAmelCase__ , self.get_expected_events(lowerCAmelCase__)) __SCREAMING_SNAKE_CASE = self.get_trainer(callbacks=[MyTestTrainerCallback] , evaluation_strategy="""epoch""") trainer.train() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowerCAmelCase__ , self.get_expected_events(lowerCAmelCase__)) # A bit of everything __SCREAMING_SNAKE_CASE = self.get_trainer( callbacks=[MyTestTrainerCallback] , logging_steps=3 , save_steps=1_0 , eval_steps=5 , evaluation_strategy="""steps""" , ) trainer.train() __SCREAMING_SNAKE_CASE = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowerCAmelCase__ , self.get_expected_events(lowerCAmelCase__)) # warning should be emitted for duplicated callbacks with patch("""transformers.trainer_callback.logger.warning""") as warn_mock: __SCREAMING_SNAKE_CASE = self.get_trainer( callbacks=[MyTestTrainerCallback, MyTestTrainerCallback] , ) assert str(lowerCAmelCase__) in warn_mock.call_args[0][0]
100
"""simple docstring""" import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( """kwargs, expected""" , [ ({"""num_shards""": 0, """max_num_jobs""": 1}, []), ({"""num_shards""": 10, """max_num_jobs""": 1}, [range(10 )]), ({"""num_shards""": 10, """max_num_jobs""": 10}, [range(UpperCamelCase_ , i + 1 ) for i in range(10 )]), ({"""num_shards""": 1, """max_num_jobs""": 10}, [range(1 )]), ({"""num_shards""": 10, """max_num_jobs""": 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({"""num_shards""": 3, """max_num_jobs""": 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = _distribute_shards(**UpperCamelCase_ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, max_num_jobs, expected""" , [ ({"""foo""": 0}, 10, [{"""foo""": 0}]), ({"""shards""": [0, 1, 2, 3]}, 1, [{"""shards""": [0, 1, 2, 3]}]), ({"""shards""": [0, 1, 2, 3]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}, {"""shards""": [2]}, {"""shards""": [3]}]), ({"""shards""": [0, 1]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}]), ({"""shards""": [0, 1, 2, 3]}, 2, [{"""shards""": [0, 1]}, {"""shards""": [2, 3]}]), ] , ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = _split_gen_kwargs(UpperCamelCase_ , UpperCamelCase_ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, expected""" , [ ({"""foo""": 0}, 1), ({"""shards""": [0]}, 1), ({"""shards""": [0, 1, 2, 3]}, 4), ({"""shards""": [0, 1, 2, 3], """foo""": 0}, 4), ({"""shards""": [0, 1, 2, 3], """other""": (0, 1)}, 4), ({"""shards""": [0, 1, 2, 3], """shards2""": [0, 1]}, RuntimeError), ] , ) def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): if expected is RuntimeError: with pytest.raises(UpperCamelCase_ ): _number_of_shards_in_gen_kwargs(UpperCamelCase_ ) else: __SCREAMING_SNAKE_CASE = _number_of_shards_in_gen_kwargs(UpperCamelCase_ ) assert out == expected
100
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = {"""configuration_reformer""": ["""REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ReformerConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ["""ReformerTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ["""ReformerTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ """REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """ReformerAttention""", """ReformerForMaskedLM""", """ReformerForQuestionAnswering""", """ReformerForSequenceClassification""", """ReformerLayer""", """ReformerModel""", """ReformerModelWithLMHead""", """ReformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_reformer import REFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ReformerConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer import ReformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_reformer_fast import ReformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_reformer import ( REFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ReformerAttention, ReformerForMaskedLM, ReformerForQuestionAnswering, ReformerForSequenceClassification, ReformerLayer, ReformerModel, ReformerModelWithLMHead, ReformerPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
353
from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE_ ( __lowerCAmelCase ): def __init__( self : Dict , lowerCamelCase_ : Tuple , lowerCamelCase_ : Optional[Any] ): """simple docstring""" super().__init__() # make sure scheduler can always be converted to DDIM UpperCamelCase = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=lowerCamelCase_ , scheduler=lowerCamelCase_ ) @torch.no_grad() def __call__( self : Union[str, Any] , lowerCamelCase_ : int = 1 , lowerCamelCase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase_ : float = 0.0 , lowerCamelCase_ : int = 50 , lowerCamelCase_ : Optional[bool] = None , lowerCamelCase_ : Optional[str] = "pil" , lowerCamelCase_ : bool = True , ): """simple docstring""" if isinstance(self.unet.config.sample_size , lowerCamelCase_ ): UpperCamelCase = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: UpperCamelCase = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) and len(lowerCamelCase_ ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(lowerCamelCase_ )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) UpperCamelCase = randn_tensor(lowerCamelCase_ , generator=lowerCamelCase_ , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(lowerCamelCase_ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output UpperCamelCase = self.unet(lowerCamelCase_ , lowerCamelCase_ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 UpperCamelCase = self.scheduler.step( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , eta=lowerCamelCase_ , use_clipped_model_output=lowerCamelCase_ , generator=lowerCamelCase_ ).prev_sample UpperCamelCase = (image / 2 + 0.5).clamp(0 , 1 ) UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCamelCase = self.numpy_to_pil(lowerCamelCase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase_ )
165
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { "xlm-mlm-en-2048": "https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json", "xlm-mlm-ende-1024": "https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json", "xlm-mlm-enfr-1024": "https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json", "xlm-mlm-enro-1024": "https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json", "xlm-mlm-tlm-xnli15-1024": "https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json", "xlm-mlm-xnli15-1024": "https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json", "xlm-clm-enfr-1024": "https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json", "xlm-clm-ende-1024": "https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json", "xlm-mlm-17-1280": "https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json", "xlm-mlm-100-1280": "https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json", } class UpperCAmelCase_ ( _a ): """simple docstring""" lowercase = "xlm" lowercase = { "hidden_size": "emb_dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", "n_words": "vocab_size", # For backward compatibility } def __init__( self : Tuple , snake_case_ : Optional[int]=30_145 , snake_case_ : Optional[Any]=2_048 , snake_case_ : Optional[Any]=12 , snake_case_ : Dict=16 , snake_case_ : Optional[Any]=0.1 , snake_case_ : str=0.1 , snake_case_ : Tuple=True , snake_case_ : Union[str, Any]=False , snake_case_ : Dict=False , snake_case_ : Optional[int]=False , snake_case_ : Dict=1 , snake_case_ : List[Any]=True , snake_case_ : Tuple=512 , snake_case_ : List[str]=2_048**-0.5 , snake_case_ : List[Any]=1E-1_2 , snake_case_ : int=0.02 , snake_case_ : List[str]=0 , snake_case_ : Optional[int]=1 , snake_case_ : List[Any]=2 , snake_case_ : str=3 , snake_case_ : Union[str, Any]=5 , snake_case_ : List[str]=True , snake_case_ : List[str]="first" , snake_case_ : List[Any]=True , snake_case_ : List[Any]=None , snake_case_ : List[str]=True , snake_case_ : Dict=0.1 , snake_case_ : Any=5 , snake_case_ : Optional[Any]=5 , snake_case_ : List[Any]=0 , snake_case_ : List[str]=0 , snake_case_ : str=2 , snake_case_ : Tuple=0 , **snake_case_ : List[Any] , ): snake_case__ : Optional[Any] = vocab_size snake_case__ : List[Any] = emb_dim snake_case__ : int = n_layers snake_case__ : Dict = n_heads snake_case__ : Union[str, Any] = dropout snake_case__ : List[str] = attention_dropout snake_case__ : Optional[int] = gelu_activation snake_case__ : Union[str, Any] = sinusoidal_embeddings snake_case__ : Union[str, Any] = causal snake_case__ : Tuple = asm snake_case__ : int = n_langs snake_case__ : int = use_lang_emb snake_case__ : List[Any] = layer_norm_eps snake_case__ : Union[str, Any] = bos_index snake_case__ : int = eos_index snake_case__ : str = pad_index snake_case__ : str = unk_index snake_case__ : Tuple = mask_index snake_case__ : Optional[int] = is_encoder snake_case__ : int = max_position_embeddings snake_case__ : List[Any] = embed_init_std snake_case__ : List[str] = init_std snake_case__ : Any = summary_type snake_case__ : Tuple = summary_use_proj snake_case__ : int = summary_activation snake_case__ : Optional[int] = summary_proj_to_labels snake_case__ : Optional[int] = summary_first_dropout snake_case__ : str = start_n_top snake_case__ : Union[str, Any] = end_n_top snake_case__ : List[Any] = mask_token_id snake_case__ : Optional[Any] = lang_id if "n_words" in kwargs: snake_case__ : Any = kwargs["""n_words"""] super().__init__(pad_token_id=snake_case_ , bos_token_id=snake_case_ , **snake_case_ ) class UpperCAmelCase_ ( _a ): """simple docstring""" @property def lowerCamelCase ( self : Dict ): if self.task == "multiple-choice": snake_case__ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: snake_case__ : Optional[int] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ("""token_type_ids""", dynamic_axis), ] )
35
'''simple docstring''' import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def __snake_case( _lowerCAmelCase ) -> Optional[int]: snake_case__ : int = FileLock(str(tmpdir / """foo.lock""" ) ) snake_case__ : Dict = FileLock(str(tmpdir / """foo.lock""" ) ) snake_case__ : List[str] = 0.01 with locka.acquire(): with pytest.raises(_lowerCAmelCase ): snake_case__ : str = time.time() locka.acquire(_lowerCAmelCase ) assert time.time() - _start > timeout def __snake_case( _lowerCAmelCase ) -> Tuple: snake_case__ : Dict = """a""" * 1_000 + """.lock""" snake_case__ : int = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(""".lock""" ) assert not locka._lock_file.endswith(_lowerCAmelCase ) assert len(os.path.basename(locka._lock_file ) ) <= 255 snake_case__ : Dict = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(_lowerCAmelCase ): locka.acquire(0 )
35
1
'''simple docstring''' from __future__ import annotations def __lowerCamelCase ( A__ ) -> int: """simple docstring""" UpperCamelCase = len(A__ ) // 2 # choose the middle 3 elements UpperCamelCase = lst[m - 1 : m + 2] # if middle element is peak if three[1] > three[0] and three[1] > three[2]: return three[1] # if increasing, recurse on right elif three[0] < three[2]: if len(lst[:m] ) == 2: m -= 1 return peak(lst[m:] ) # decreasing else: if len(lst[:m] ) == 2: m += 1 return peak(lst[:m] ) if __name__ == "__main__": import doctest doctest.testmod()
355
'''simple docstring''' import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = (UnCLIPScheduler,) def A ( self : Union[str, Any] , **UpperCamelCase__ : Any ): """simple docstring""" UpperCamelCase = { 'num_train_timesteps': 1_0_0_0, 'variance_type': 'fixed_small_log', 'clip_sample': True, 'clip_sample_range': 1.0, 'prediction_type': 'epsilon', } config.update(**UpperCamelCase__ ) return config def A ( self : str ): """simple docstring""" for timesteps in [1, 5, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=UpperCamelCase__ ) def A ( self : List[str] ): """simple docstring""" for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=UpperCamelCase__ ) def A ( self : List[Any] ): """simple docstring""" for clip_sample in [True, False]: self.check_over_configs(clip_sample=UpperCamelCase__ ) def A ( self : Optional[int] ): """simple docstring""" for clip_sample_range in [1, 5, 1_0, 2_0]: self.check_over_configs(clip_sample_range=UpperCamelCase__ ) def A ( self : Tuple ): """simple docstring""" for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=UpperCamelCase__ ) def A ( self : Union[str, Any] ): """simple docstring""" for time_step in [0, 5_0_0, 9_9_9]: for prev_timestep in [None, 5, 1_0_0, 2_5_0, 5_0_0, 7_5_0]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=UpperCamelCase__ , prev_timestep=UpperCamelCase__ ) def A ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config(variance_type='fixed_small_log' ) UpperCamelCase = scheduler_class(**UpperCamelCase__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0_0_0_0E-1_0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_8_7 ) - 0.0_5_4_9_6_2_5 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_9_9 ) - 0.9_9_9_4_9_8_7 ) ) < 1E-5 def A ( self : Tuple ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config(variance_type='learned_range' ) UpperCamelCase = scheduler_class(**UpperCamelCase__ ) UpperCamelCase = 0.5 assert scheduler._get_variance(1 , predicted_variance=UpperCamelCase__ ) - -1_0.1_7_1_2_7_9_0 < 1E-5 assert scheduler._get_variance(4_8_7 , predicted_variance=UpperCamelCase__ ) - -5.7_9_9_8_0_5_2 < 1E-5 assert scheduler._get_variance(9_9_9 , predicted_variance=UpperCamelCase__ ) - -0.0_0_1_0_0_1_1 < 1E-5 def A ( self : int ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config() UpperCamelCase = scheduler_class(**UpperCamelCase__ ) UpperCamelCase = scheduler.timesteps UpperCamelCase = self.dummy_model() UpperCamelCase = self.dummy_sample_deter UpperCamelCase = torch.manual_seed(0 ) for i, t in enumerate(UpperCamelCase__ ): # 1. predict noise residual UpperCamelCase = model(UpperCamelCase__ , UpperCamelCase__ ) # 2. predict previous mean of sample x_t-1 UpperCamelCase = scheduler.step(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample UpperCamelCase = pred_prev_sample UpperCamelCase = torch.sum(torch.abs(UpperCamelCase__ ) ) UpperCamelCase = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 2_5_2.2_6_8_2_4_9_5 ) < 1E-2 assert abs(result_mean.item() - 0.3_2_8_4_7_4_3 ) < 1E-3 def A ( self : Union[str, Any] ): """simple docstring""" UpperCamelCase = self.scheduler_classes[0] UpperCamelCase = self.get_scheduler_config() UpperCamelCase = scheduler_class(**UpperCamelCase__ ) scheduler.set_timesteps(2_5 ) UpperCamelCase = scheduler.timesteps UpperCamelCase = self.dummy_model() UpperCamelCase = self.dummy_sample_deter UpperCamelCase = torch.manual_seed(0 ) for i, t in enumerate(UpperCamelCase__ ): # 1. predict noise residual UpperCamelCase = model(UpperCamelCase__ , UpperCamelCase__ ) if i + 1 == timesteps.shape[0]: UpperCamelCase = None else: UpperCamelCase = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 UpperCamelCase = scheduler.step( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , prev_timestep=UpperCamelCase__ , generator=UpperCamelCase__ ).prev_sample UpperCamelCase = pred_prev_sample UpperCamelCase = torch.sum(torch.abs(UpperCamelCase__ ) ) UpperCamelCase = torch.mean(torch.abs(UpperCamelCase__ ) ) assert abs(result_sum.item() - 2_5_8.2_0_4_4_9_8_3 ) < 1E-2 assert abs(result_mean.item() - 0.3_3_6_2_0_3_8 ) < 1E-3 def A ( self : Tuple ): """simple docstring""" pass def A ( self : Optional[int] ): """simple docstring""" pass
249
0
"""simple docstring""" import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm __magic_name__ = re.compile("[^A-Za-z_0-9]") # parameters used in DuplicationIndex __magic_name__ = 10 __magic_name__ = 256 def _lowerCAmelCase ( UpperCamelCase_ ): if len(_A ) < MIN_NUM_TOKENS: return None __SCREAMING_SNAKE_CASE = MinHash(num_perm=_A ) for token in set(_A ): min_hash.update(token.encode() ) return min_hash def _lowerCAmelCase ( UpperCamelCase_ ): return {t for t in NON_ALPHA.split(_A ) if len(t.strip() ) > 0} class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , *, lowerCAmelCase__ = 0.85 , ): __SCREAMING_SNAKE_CASE = duplication_jaccard_threshold __SCREAMING_SNAKE_CASE = NUM_PERM __SCREAMING_SNAKE_CASE = MinHashLSH(threshold=self._duplication_jaccard_threshold , num_perm=self._num_perm) __SCREAMING_SNAKE_CASE = defaultdict(lowerCAmelCase__) def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self._index.query(lowerCAmelCase__) if code_key in self._index.keys: print(f"Duplicate key {code_key}") return self._index.insert(lowerCAmelCase__ , lowerCAmelCase__) if len(lowerCAmelCase__) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(lowerCAmelCase__) break else: self._duplicate_clusters[close_duplicates[0]].add(lowerCAmelCase__) def snake_case_ ( self): __SCREAMING_SNAKE_CASE = [] for base, duplicates in self._duplicate_clusters.items(): __SCREAMING_SNAKE_CASE = [base] + list(lowerCAmelCase__) # reformat the cluster to be a list of dict __SCREAMING_SNAKE_CASE = [{"""base_index""": el[0], """repo_name""": el[1], """path""": el[2]} for el in cluster] duplicate_clusters.append(lowerCAmelCase__) return duplicate_clusters def snake_case_ ( self , lowerCAmelCase__): __SCREAMING_SNAKE_CASE = self.get_duplicate_clusters() with open(lowerCAmelCase__ , """w""") as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__) def _lowerCAmelCase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = element __SCREAMING_SNAKE_CASE = get_min_hash([t for t in NON_ALPHA.split(data["""content"""] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def _lowerCAmelCase ( UpperCamelCase_ ): with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(_A , max_queue_size=1_0000 ) , chunksize=100 , ): if data is not None: yield data def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = DuplicationIndex(duplication_jaccard_threshold=_A ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(_A ) ) , max_queue_size=100 ) ): di.add(_A , _A ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = get_tokens(_A ) __SCREAMING_SNAKE_CASE = get_tokens(_A ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) __magic_name__ = None def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ ): __SCREAMING_SNAKE_CASE = [] for elementa in cluster: __SCREAMING_SNAKE_CASE = _shared_dataset[elementa["""base_index"""]]["""content"""] for elementa in extremes: __SCREAMING_SNAKE_CASE = _shared_dataset[elementa["""base_index"""]]["""content"""] if jaccard_similarity(_A , _A ) >= jaccard_threshold: elementa["copies"] += 1 break else: __SCREAMING_SNAKE_CASE = 1 extremes.append(_A ) return extremes def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): global _shared_dataset __SCREAMING_SNAKE_CASE = dataset __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = partial(_find_cluster_extremes_shared , jaccard_threshold=_A ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( _A , _A , ) , total=len(_A ) , ): extremes_list.append(_A ) return extremes_list def _lowerCAmelCase ( UpperCamelCase_ , UpperCamelCase_ = 0.85 ): __SCREAMING_SNAKE_CASE = make_duplicate_clusters(_A , _A ) __SCREAMING_SNAKE_CASE = {x["""base_index"""] for cluster in duplicate_clusters for x in cluster} __SCREAMING_SNAKE_CASE = {} __SCREAMING_SNAKE_CASE = find_extremes(_A , _A , _A ) for extremes in extremes_clusters: for element in extremes: __SCREAMING_SNAKE_CASE = element __SCREAMING_SNAKE_CASE = duplicate_indices - set(extreme_dict.keys() ) __SCREAMING_SNAKE_CASE = dataset.filter(lambda UpperCamelCase_ , UpperCamelCase_ : idx not in remove_indices , with_indices=_A ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: __SCREAMING_SNAKE_CASE = element["""base_index"""] in extreme_dict if element["is_extreme"]: __SCREAMING_SNAKE_CASE = extreme_dict[element["""base_index"""]]["""copies"""] print(f"Original dataset size: {len(_A )}" ) print(f"Number of duplicate clusters: {len(_A )}" ) print(f"Files in duplicate cluster: {len(_A )}" ) print(f"Unique files in duplicate cluster: {len(_A )}" ) print(f"Filtered dataset size: {len(_A )}" ) return ds_filter, duplicate_clusters
100
from __future__ import annotations import math import random from collections.abc import Collection from typing import overload class snake_case__ : def __init__( self , lowerCAmelCase__ = None ) -> None: if components is None: __magic_name__ : Any = [] __magic_name__ : List[str] = list(lowerCAmelCase__ ) def __len__( self ) -> int: return len(self.__components ) def __str__( self ) -> str: return "(" + ",".join(map(lowerCAmelCase__ , self.__components ) ) + ")" def __add__( self , lowerCAmelCase__ ) -> Vector: __magic_name__ : Dict = len(self ) if size == len(lowerCAmelCase__ ): __magic_name__ : str = [self.__components[i] + other.component(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ )] return Vector(lowerCAmelCase__ ) else: raise Exception("""must have the same size""" ) def __sub__( self , lowerCAmelCase__ ) -> Vector: __magic_name__ : int = len(self ) if size == len(lowerCAmelCase__ ): __magic_name__ : str = [self.__components[i] - other.component(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ )] return Vector(lowerCAmelCase__ ) else: # error case raise Exception("""must have the same size""" ) @overload def __mul__( self , lowerCAmelCase__ ) -> Vector: ... @overload def __mul__( self , lowerCAmelCase__ ) -> float: ... def __mul__( self , lowerCAmelCase__ ) -> float | Vector: if isinstance(lowerCAmelCase__ , (float, int) ): __magic_name__ : Optional[Any] = [c * other for c in self.__components] return Vector(lowerCAmelCase__ ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and len(self ) == len(lowerCAmelCase__ ): __magic_name__ : Optional[Any] = len(self ) __magic_name__ : List[Any] = [self.__components[i] * other.component(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ )] return sum(lowerCAmelCase__ ) else: # error case raise Exception("""invalid operand!""" ) def __magic_name__ ( self ) -> Vector: return Vector(self.__components ) def __magic_name__ ( self , lowerCAmelCase__ ) -> float: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and -len(self.__components ) <= i < len(self.__components ): return self.__components[i] else: raise Exception("""index out of range""" ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: assert -len(self.__components ) <= pos < len(self.__components ) __magic_name__ : Optional[int] = value def __magic_name__ ( self ) -> float: if len(self.__components ) == 0: raise Exception("""Vector is empty""" ) __magic_name__ : Dict = [c**2 for c in self.__components] return math.sqrt(sum(lowerCAmelCase__ ) ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = False ) -> float: __magic_name__ : Optional[Any] = self * other __magic_name__ : List[str] = self.euclidean_length() * other.euclidean_length() if deg: return math.degrees(math.acos(num / den ) ) else: return math.acos(num / den ) def UpperCamelCase ( _A ): """simple docstring""" assert isinstance(_A, _A ) return Vector([0] * dimension ) def UpperCamelCase ( _A, _A ): """simple docstring""" assert isinstance(_A, _A ) and (isinstance(_A, _A )) __magic_name__ : Union[str, Any] = [0] * dimension __magic_name__ : Optional[int] = 1 return Vector(_A ) def UpperCamelCase ( _A, _A, _A ): """simple docstring""" assert ( isinstance(_A, _A ) and isinstance(_A, _A ) and (isinstance(_A, (int, float) )) ) return x * scalar + y def UpperCamelCase ( _A, _A, _A ): """simple docstring""" random.seed(_A ) __magic_name__ : Union[str, Any] = [random.randint(_A, _A ) for _ in range(_A )] return Vector(_A ) class snake_case__ : def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: __magic_name__ : Dict = matrix __magic_name__ : Tuple = w __magic_name__ : Union[str, Any] = h def __str__( self ) -> str: __magic_name__ : Dict = """""" for i in range(self.__height ): ans += "|" for j in range(self.__width ): if j < self.__width - 1: ans += str(self.__matrix[i][j] ) + "," else: ans += str(self.__matrix[i][j] ) + "|\n" return ans def __add__( self , lowerCAmelCase__ ) -> Matrix: if self.__width == other.width() and self.__height == other.height(): __magic_name__ : Tuple = [] for i in range(self.__height ): __magic_name__ : Tuple = [ self.__matrix[i][j] + other.component(lowerCAmelCase__ , lowerCAmelCase__ ) for j in range(self.__width ) ] matrix.append(lowerCAmelCase__ ) return Matrix(lowerCAmelCase__ , self.__width , self.__height ) else: raise Exception("""matrix must have the same dimension!""" ) def __sub__( self , lowerCAmelCase__ ) -> Matrix: if self.__width == other.width() and self.__height == other.height(): __magic_name__ : Optional[Any] = [] for i in range(self.__height ): __magic_name__ : int = [ self.__matrix[i][j] - other.component(lowerCAmelCase__ , lowerCAmelCase__ ) for j in range(self.__width ) ] matrix.append(lowerCAmelCase__ ) return Matrix(lowerCAmelCase__ , self.__width , self.__height ) else: raise Exception("""matrices must have the same dimension!""" ) @overload def __mul__( self , lowerCAmelCase__ ) -> Matrix: ... @overload def __mul__( self , lowerCAmelCase__ ) -> Vector: ... def __mul__( self , lowerCAmelCase__ ) -> Vector | Matrix: if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): # matrix-vector if len(lowerCAmelCase__ ) == self.__width: __magic_name__ : Tuple = zero_vector(self.__height ) for i in range(self.__height ): __magic_name__ : Optional[int] = [ self.__matrix[i][j] * other.component(lowerCAmelCase__ ) for j in range(self.__width ) ] ans.change_component(lowerCAmelCase__ , sum(lowerCAmelCase__ ) ) return ans else: raise Exception( """vector must have the same size as the """ """number of columns of the matrix!""" ) elif isinstance(lowerCAmelCase__ , (int, float) ): # matrix-scalar __magic_name__ : Any = [ [self.__matrix[i][j] * other for j in range(self.__width )] for i in range(self.__height ) ] return Matrix(lowerCAmelCase__ , self.__width , self.__height ) return None def __magic_name__ ( self ) -> int: return self.__height def __magic_name__ ( self ) -> int: return self.__width def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: if 0 <= x < self.__height and 0 <= y < self.__width: return self.__matrix[x][y] else: raise Exception("""change_component: indices out of bounds""" ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> None: if 0 <= x < self.__height and 0 <= y < self.__width: __magic_name__ : List[Any] = value else: raise Exception("""change_component: indices out of bounds""" ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: if self.__height != self.__width: raise Exception("""Matrix is not square""" ) __magic_name__ : Optional[int] = self.__matrix[:x] + self.__matrix[x + 1 :] for i in range(len(lowerCAmelCase__ ) ): __magic_name__ : List[str] = minor[i][:y] + minor[i][y + 1 :] return Matrix(lowerCAmelCase__ , self.__width - 1 , self.__height - 1 ).determinant() def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> float: if self.__height != self.__width: raise Exception("""Matrix is not square""" ) if 0 <= x < self.__height and 0 <= y < self.__width: return (-1) ** (x + y) * self.minor(lowerCAmelCase__ , lowerCAmelCase__ ) else: raise Exception("""Indices out of bounds""" ) def __magic_name__ ( self ) -> float: if self.__height != self.__width: raise Exception("""Matrix is not square""" ) if self.__height < 1: raise Exception("""Matrix has no element""" ) elif self.__height == 1: return self.__matrix[0][0] elif self.__height == 2: return ( self.__matrix[0][0] * self.__matrix[1][1] - self.__matrix[0][1] * self.__matrix[1][0] ) else: __magic_name__ : str = [ self.__matrix[0][y] * self.cofactor(0 , lowerCAmelCase__ ) for y in range(self.__width ) ] return sum(lowerCAmelCase__ ) def UpperCamelCase ( _A ): """simple docstring""" __magic_name__ : list[list[float]] = [[0] * n for _ in range(_A )] return Matrix(_A, _A, _A ) def UpperCamelCase ( _A, _A, _A, _A ): """simple docstring""" random.seed(_A ) __magic_name__ : list[list[float]] = [ [random.randint(_A, _A ) for _ in range(_A )] for _ in range(_A ) ] return Matrix(_A, _A, _A )
342
0
import argparse from pathlib import Path import torch from packaging import version from torch.onnx import export from diffusers import AutoencoderKL UpperCAmelCase_ : Union[str, Any] = version.parse(version.parse(torch.__version__).base_version) < version.parse('1.11') def SCREAMING_SNAKE_CASE_ ( __A : Dict , __A : Optional[int] , __A : Any , __A : List[str] , __A : str , __A : Optional[Any] , __A : Optional[Any] , __A : Optional[int]=False , ) -> Optional[int]: """simple docstring""" output_path.parent.mkdir(parents=A__ , exist_ok=A__ ) # PyTorch deprecated the `enable_onnx_checker` and `use_external_data_format` arguments in v1.11, # so we check the torch version for backwards compatibility if is_torch_less_than_1_11: export( A__ , A__ , f=output_path.as_posix() , input_names=A__ , output_names=A__ , dynamic_axes=A__ , do_constant_folding=A__ , use_external_data_format=A__ , enable_onnx_checker=A__ , opset_version=A__ , ) else: export( A__ , A__ , f=output_path.as_posix() , input_names=A__ , output_names=A__ , dynamic_axes=A__ , do_constant_folding=A__ , opset_version=A__ , ) @torch.no_grad() def SCREAMING_SNAKE_CASE_ ( __A : List[str] , __A : List[str] , __A : str , __A : Union[str, Any] = False ) -> List[Any]: """simple docstring""" a_ : Tuple = torch.floataa if fpaa else torch.floataa if fpaa and torch.cuda.is_available(): a_ : Tuple = 'cuda' elif fpaa and not torch.cuda.is_available(): raise ValueError('`float16` model export is only supported on GPUs with CUDA' ) else: a_ : int = 'cpu' a_ : Optional[int] = Path(A__ ) # VAE DECODER a_ : Optional[Any] = AutoencoderKL.from_pretrained(model_path + '/vae' ) a_ : Dict = vae_decoder.config.latent_channels # forward only through the decoder part a_ : Union[str, Any] = vae_decoder.decode onnx_export( A__ , model_args=( torch.randn(1 , A__ , 25 , 25 ).to(device=A__ , dtype=A__ ), False, ) , output_path=output_path / 'vae_decoder' / 'model.onnx' , ordered_input_names=['latent_sample', 'return_dict'] , output_names=['sample'] , dynamic_axes={ 'latent_sample': {0: 'batch', 1: 'channels', 2: 'height', 3: 'width'}, } , opset=A__ , ) del vae_decoder if __name__ == "__main__": UpperCAmelCase_ : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--model_path', type=str, required=True, help='Path to the `diffusers` checkpoint to convert (either a local directory or on the Hub).', ) parser.add_argument('--output_path', type=str, required=True, help='Path to the output model.') parser.add_argument( '--opset', default=14, type=int, help='The version of the ONNX operator set to use.', ) parser.add_argument('--fp16', action='store_true', default=False, help='Export the models in `float16` mode') UpperCAmelCase_ : Tuple = parser.parse_args() print(args.output_path) convert_models(args.model_path, args.output_path, args.opset, args.fpaa) print('SD: Done: ONNX')
358
import json import sys def SCREAMING_SNAKE_CASE_ ( __A : Optional[Any] , __A : List[str] ) -> Tuple: """simple docstring""" with open(__A , encoding='utf-8' ) as f: a_ : Union[str, Any] = json.load(__A ) a_ : Any = ['<details>', '<summary>Show updated benchmarks!</summary>', ' '] for benchmark_name in sorted(__A ): a_ : List[str] = results[benchmark_name] a_ : int = benchmark_name.split('/' )[-1] output_md.append(F"""### Benchmark: {benchmark_file_name}""" ) a_ : Any = '| metric |' a_ : Optional[Any] = '|--------|' a_ : int = '| new / old (diff) |' for metric_name in sorted(__A ): a_ : List[Any] = benchmark_res[metric_name] a_ : int = metric_vals['new'] a_ : Union[str, Any] = metric_vals.get('old' , __A ) a_ : Optional[int] = metric_vals.get('diff' , __A ) a_ : str = F""" {new_val:f}""" if isinstance(__A , (int, float) ) else 'None' if old_val is not None: val_str += F""" / {old_val:f}""" if isinstance(__A , (int, float) ) else "None" if dif_val is not None: val_str += F""" ({dif_val:f})""" if isinstance(__A , (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('</details>' ) with open(__A , 'w' , encoding='utf-8' ) as f: f.writelines('\n'.join(__A ) ) if __name__ == "__main__": UpperCAmelCase_ : int = sys.argv[1] UpperCAmelCase_ : Any = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
120
0
def UpperCamelCase ( _A ): """simple docstring""" if num <= 0: raise ValueError("""Input must be a positive integer""" ) __magic_name__ : int = [True] * (num + 1) __magic_name__ : Optional[int] = 2 while p * p <= num: if primes[p]: for i in range(p * p, num + 1, _A ): __magic_name__ : List[str] = False p += 1 return [prime for prime in range(2, num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() __magic_name__: Union[str, Any] = int(input("Enter a positive integer: ").strip()) print(prime_sieve_eratosthenes(user_num))
342
import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class snake_case__ ( unittest.TestCase ): def __magic_name__ ( self ) -> str: __magic_name__ : Tuple = """ylacombe/bark-small""" __magic_name__ : List[str] = tempfile.mkdtemp() __magic_name__ : Optional[Any] = """en_speaker_1""" __magic_name__ : Union[str, Any] = """This is a test string""" __magic_name__ : Optional[int] = """speaker_embeddings_path.json""" __magic_name__ : Any = """speaker_embeddings""" def __magic_name__ ( self , **lowerCAmelCase__ ) -> List[Any]: return AutoTokenizer.from_pretrained(self.checkpoint , **lowerCAmelCase__ ) def __magic_name__ ( self ) -> Optional[Any]: shutil.rmtree(self.tmpdirname ) def __magic_name__ ( self ) -> Tuple: __magic_name__ : Optional[Any] = self.get_tokenizer() __magic_name__ : int = BarkProcessor(tokenizer=lowerCAmelCase__ ) processor.save_pretrained(self.tmpdirname ) __magic_name__ : Union[str, Any] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def __magic_name__ ( self ) -> Optional[int]: __magic_name__ : Optional[int] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) __magic_name__ : Optional[Any] = self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) __magic_name__ : str = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="""(BOS)""" , eos_token="""(EOS)""" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def __magic_name__ ( self ) -> Any: __magic_name__ : List[str] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) __magic_name__ : Union[str, Any] = 35 __magic_name__ : List[Any] = 2 __magic_name__ : Dict = 8 __magic_name__ : Tuple = { """semantic_prompt""": np.ones(lowerCAmelCase__ ), """coarse_prompt""": np.ones((nb_codebooks_coarse, seq_len) ), """fine_prompt""": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset __magic_name__ : Optional[int] = processor(text=self.input_string , voice_preset=lowerCAmelCase__ ) __magic_name__ : Union[str, Any] = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from npz file __magic_name__ : Dict = os.path.join(self.tmpdirname , """file.npz""" ) np.savez(lowerCAmelCase__ , **lowerCAmelCase__ ) __magic_name__ : Optional[Any] = processor(text=self.input_string , voice_preset=lowerCAmelCase__ ) __magic_name__ : List[Any] = inputs["""history_prompt"""] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowerCAmelCase__ , np.array([] ) ).tolist() ) # test loading voice preset from the hub __magic_name__ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def __magic_name__ ( self ) -> Optional[Any]: __magic_name__ : str = self.get_tokenizer() __magic_name__ : Dict = BarkProcessor(tokenizer=lowerCAmelCase__ ) __magic_name__ : Optional[Any] = processor(text=self.input_string ) __magic_name__ : List[Any] = tokenizer( self.input_string , padding="""max_length""" , max_length=2_56 , add_special_tokens=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ , return_token_type_ids=lowerCAmelCase__ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
342
1
"""simple docstring""" from __future__ import annotations from collections.abc import Sequence from typing import Literal def lowerCamelCase (a_ :str , a_ :str) -> str | Literal[False]: lowercase :Union[str, Any] = list(a_) lowercase :Optional[Any] = list(a_) lowercase :str = 0 for i in range(len(a_)): if lista[i] != lista[i]: count += 1 lowercase :str = '''_''' if count > 1: return False else: return "".join(a_) def lowerCamelCase (a_ :list[str]) -> list[str]: lowercase :Optional[Any] = [] while True: lowercase :Tuple = ['''$'''] * len(a_) lowercase :Tuple = [] for i in range(len(a_)): for j in range(i + 1 , len(a_)): lowercase :Optional[int] = compare_string(binary[i] , binary[j]) if k is False: lowercase :Tuple = '''*''' lowercase :Any = '''*''' temp.append('''X''') for i in range(len(a_)): if checka[i] == "$": pi.append(binary[i]) if len(a_) == 0: return pi lowercase :str = list(set(a_)) def lowerCamelCase (a_ :int , a_ :Sequence[float]) -> list[str]: lowercase :Optional[int] = [] for minterm in minterms: lowercase :List[str] = '''''' for _ in range(a_): lowercase :List[str] = str(minterm % 2) + string minterm //= 2 temp.append(a_) return temp def lowerCamelCase (a_ :str , a_ :str , a_ :int) -> bool: lowercase :int = list(a_) lowercase :str = list(a_) lowercase :List[str] = 0 for i in range(len(a_)): if lista[i] != lista[i]: count_n += 1 return count_n == count def lowerCamelCase (a_ :list[list[int]] , a_ :list[str]) -> list[str]: lowercase :Any = [] lowercase :List[Any] = [0] * len(a_) for i in range(len(chart[0])): lowercase :List[Any] = 0 lowercase :int = -1 for j in range(len(a_)): if chart[j][i] == 1: count += 1 lowercase :List[Any] = j if count == 1: lowercase :Tuple = 1 for i in range(len(a_)): if select[i] == 1: for j in range(len(chart[0])): if chart[i][j] == 1: for k in range(len(a_)): lowercase :List[str] = 0 temp.append(prime_implicants[i]) while True: lowercase :Tuple = 0 lowercase :Dict = -1 lowercase :int = 0 for i in range(len(a_)): lowercase :List[Any] = chart[i].count(1) if count_n > max_n: lowercase :List[Any] = count_n lowercase :int = i if max_n == 0: return temp temp.append(prime_implicants[rem]) for i in range(len(chart[0])): if chart[rem][i] == 1: for j in range(len(a_)): lowercase :Tuple = 0 def lowerCamelCase (a_ :list[str] , a_ :list[str]) -> list[list[int]]: lowercase :Dict = [[0 for x in range(len(a_))] for x in range(len(a_))] for i in range(len(a_)): lowercase :Any = prime_implicants[i].count('''_''') for j in range(len(a_)): if is_for_table(prime_implicants[i] , binary[j] , a_): lowercase :int = 1 return chart def lowerCamelCase () -> None: lowercase :int = int(input('''Enter the no. of variables\n''')) lowercase :Tuple = [ float(a_) for x in input( '''Enter the decimal representation of Minterms \'Spaces Separated\'\n''').split() ] lowercase :Dict = decimal_to_binary(a_ , a_) lowercase :List[Any] = check(a_) print('''Prime Implicants are:''') print(a_) lowercase :Union[str, Any] = prime_implicant_chart(a_ , a_) lowercase :Dict = selection(a_ , a_) print('''Essential Prime Implicants are:''') print(a_) if __name__ == "__main__": import doctest doctest.testmod() main()
358
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''caidas/swin2sr-classicalsr-x2-64''': ( '''https://huggingface.co/caidas/swin2sr-classicalsr-x2-64/resolve/main/config.json''' ), } class __magic_name__ ( __UpperCAmelCase ): __A : Tuple = "swin2sr" __A : Dict = { "hidden_size": "embed_dim", "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self : List[str] , snake_case__ : List[str]=6_4 , snake_case__ : Union[str, Any]=1 , snake_case__ : Tuple=3 , snake_case__ : int=1_8_0 , snake_case__ : Union[str, Any]=[6, 6, 6, 6, 6, 6] , snake_case__ : List[str]=[6, 6, 6, 6, 6, 6] , snake_case__ : Tuple=8 , snake_case__ : List[Any]=2.0 , snake_case__ : Any=True , snake_case__ : Dict=0.0 , snake_case__ : Dict=0.0 , snake_case__ : Dict=0.1 , snake_case__ : Dict="gelu" , snake_case__ : Optional[int]=False , snake_case__ : Any=0.02 , snake_case__ : Any=1e-5 , snake_case__ : Optional[int]=2 , snake_case__ : Optional[int]=1.0 , snake_case__ : Optional[Any]="1conv" , snake_case__ : List[str]="pixelshuffle" , **snake_case__ : Tuple , ): '''simple docstring''' super().__init__(**snake_case__ ) lowercase :Dict = image_size lowercase :List[str] = patch_size lowercase :Tuple = num_channels lowercase :int = embed_dim lowercase :Any = depths lowercase :Union[str, Any] = len(snake_case__ ) lowercase :List[str] = num_heads lowercase :int = window_size lowercase :Tuple = mlp_ratio lowercase :List[Any] = qkv_bias lowercase :Optional[int] = hidden_dropout_prob lowercase :Tuple = attention_probs_dropout_prob lowercase :Tuple = drop_path_rate lowercase :Optional[Any] = hidden_act lowercase :Union[str, Any] = use_absolute_embeddings lowercase :Dict = layer_norm_eps lowercase :Optional[Any] = initializer_range lowercase :Optional[Any] = upscale lowercase :Any = img_range lowercase :Optional[int] = resi_connection lowercase :Union[str, Any] = upsampler
172
0
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class UpperCAmelCase_ ( UpperCamelCase , UpperCamelCase , UpperCamelCase ): '''simple docstring''' @register_to_config def __init__( self , __A , __A , __A , __A , __A , __A , __A , __A , __A , __A = False , ): """simple docstring""" super().__init__() lowerCamelCase : Optional[Any] = nn.Embedding(__A , __A ) lowerCamelCase : Tuple = nn.Embedding(__A , __A ) lowerCamelCase : Union[str, Any] = False lowerCamelCase : Any = nn.Dropout(p=__A ) lowerCamelCase : int = TaConfig( vocab_size=__A , d_model=__A , num_heads=__A , d_kv=__A , d_ff=__A , dropout_rate=__A , feed_forward_proj=__A , is_decoder=__A , is_encoder_decoder=__A , ) lowerCamelCase : Optional[int] = nn.ModuleList() for lyr_num in range(__A ): lowerCamelCase : str = TaBlock(__A ) self.encoders.append(__A ) lowerCamelCase : str = TaLayerNorm(__A ) lowerCamelCase : Tuple = nn.Dropout(p=__A ) def _snake_case ( self , __A , __A ): """simple docstring""" lowerCamelCase : List[Any] = self.token_embedder(__A ) lowerCamelCase : Optional[int] = encoder_input_tokens.shape[1] lowerCamelCase : List[Any] = torch.arange(__A , device=encoder_input_tokens.device ) x += self.position_encoding(__A ) lowerCamelCase : int = self.dropout_pre(__A ) # inverted the attention mask lowerCamelCase : Any = encoder_input_tokens.size() lowerCamelCase : Optional[Any] = self.get_extended_attention_mask(__A , __A ) for lyr in self.encoders: lowerCamelCase : str = lyr(__A , __A )[0] lowerCamelCase : Dict = self.layer_norm(__A ) return self.dropout_post(__A ), encoder_inputs_mask
283
import math from typing import Callable, List, Optional, Union import numpy as np import PIL import torch from PIL import Image from transformers import CLIPTextModel, CLIPTokenizer from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from diffusers.schedulers import DDIMScheduler, DDPMScheduler, LMSDiscreteScheduler, PNDMScheduler def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=[] ): '''simple docstring''' lowerCamelCase : Optional[Any] = size[0] - overlap_pixels * 2 lowerCamelCase : int = size[1] - overlap_pixels * 2 for letter in ["l", "r"]: if letter in remove_borders: size_x += overlap_pixels for letter in ["t", "b"]: if letter in remove_borders: size_y += overlap_pixels lowerCamelCase : Tuple = np.ones((size_y, size_x) , dtype=np.uinta ) * 255 lowerCamelCase : List[Any] = np.pad(SCREAMING_SNAKE_CASE_ , mode="linear_ramp" , pad_width=SCREAMING_SNAKE_CASE_ , end_values=0 ) if "l" in remove_borders: lowerCamelCase : Optional[Any] = mask[:, overlap_pixels : mask.shape[1]] if "r" in remove_borders: lowerCamelCase : List[Any] = mask[:, 0 : mask.shape[1] - overlap_pixels] if "t" in remove_borders: lowerCamelCase : List[Any] = mask[overlap_pixels : mask.shape[0], :] if "b" in remove_borders: lowerCamelCase : Tuple = mask[0 : mask.shape[0] - overlap_pixels, :] return mask def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' return max(SCREAMING_SNAKE_CASE_ , min(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' return ( clamp(rect[0] , min[0] , max[0] ), clamp(rect[1] , min[1] , max[1] ), clamp(rect[2] , min[0] , max[0] ), clamp(rect[3] , min[1] , max[1] ), ) def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : Optional[Any] = list(SCREAMING_SNAKE_CASE_ ) rect[0] -= overlap rect[1] -= overlap rect[2] += overlap rect[3] += overlap lowerCamelCase : Any = clamp_rect(SCREAMING_SNAKE_CASE_ , [0, 0] , [image_size[0], image_size[1]] ) return rect def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : Dict = Image.new("RGB" , (tile.size[0] + original_slice, tile.size[1]) ) result.paste( original_image.resize((tile.size[0], tile.size[1]) , Image.BICUBIC ).crop( (slice_x, 0, slice_x + original_slice, tile.size[1]) ) , (0, 0) , ) result.paste(SCREAMING_SNAKE_CASE_ , (original_slice, 0) ) return result def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : Union[str, Any] = (original_image_slice * 4, 0, tile.size[0], tile.size[1]) lowerCamelCase : int = tile.crop(SCREAMING_SNAKE_CASE_ ) return tile def lowercase_( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): '''simple docstring''' lowerCamelCase : int = n % d return n - divisor class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' def __init__( self , __A , __A , __A , __A , __A , __A , __A = 350 , ): """simple docstring""" super().__init__( vae=__A , text_encoder=__A , tokenizer=__A , unet=__A , low_res_scheduler=__A , scheduler=__A , max_noise_level=__A , ) def _snake_case ( self , __A , __A , __A , __A , __A , __A , __A , **__A ): """simple docstring""" torch.manual_seed(0 ) lowerCamelCase : Tuple = ( min(image.size[0] - (tile_size + original_image_slice) , x * tile_size ), min(image.size[1] - (tile_size + original_image_slice) , y * tile_size ), min(image.size[0] , (x + 1) * tile_size ), min(image.size[1] , (y + 1) * tile_size ), ) lowerCamelCase : Union[str, Any] = add_overlap_rect(__A , __A , image.size ) lowerCamelCase : List[str] = image.crop(__A ) lowerCamelCase : Optional[int] = ((crop_rect[0] + ((crop_rect[2] - crop_rect[0]) / 2)) / image.size[0]) * tile.size[0] lowerCamelCase : int = translated_slice_x - (original_image_slice / 2) lowerCamelCase : Optional[Any] = max(0 , __A ) lowerCamelCase : Tuple = squeeze_tile(__A , __A , __A , __A ) lowerCamelCase : Dict = to_input.size lowerCamelCase : Optional[int] = to_input.resize((tile_size, tile_size) , Image.BICUBIC ) lowerCamelCase : Dict = super(__A , self ).__call__(image=__A , **__A ).images[0] lowerCamelCase : Tuple = upscaled_tile.resize((orig_input_size[0] * 4, orig_input_size[1] * 4) , Image.BICUBIC ) lowerCamelCase : Optional[Any] = unsqueeze_tile(__A , __A ) lowerCamelCase : Optional[Any] = upscaled_tile.resize((tile.size[0] * 4, tile.size[1] * 4) , Image.BICUBIC ) lowerCamelCase : int = [] if x == 0: remove_borders.append("l" ) elif crop_rect[2] == image.size[0]: remove_borders.append("r" ) if y == 0: remove_borders.append("t" ) elif crop_rect[3] == image.size[1]: remove_borders.append("b" ) lowerCamelCase : int = Image.fromarray( make_transparency_mask( (upscaled_tile.size[0], upscaled_tile.size[1]) , tile_border * 4 , remove_borders=__A ) , mode="L" , ) final_image.paste( __A , (crop_rect_with_overlap[0] * 4, crop_rect_with_overlap[1] * 4) , __A ) @torch.no_grad() def __call__( self , __A , __A , __A = 75 , __A = 9.0 , __A = 50 , __A = None , __A = 1 , __A = 0.0 , __A = None , __A = None , __A = None , __A = 1 , __A = 128 , __A = 32 , __A = 32 , ): """simple docstring""" lowerCamelCase : Dict = Image.new("RGB" , (image.size[0] * 4, image.size[1] * 4) ) lowerCamelCase : Union[str, Any] = math.ceil(image.size[0] / tile_size ) lowerCamelCase : Dict = math.ceil(image.size[1] / tile_size ) lowerCamelCase : str = tcx * tcy lowerCamelCase : int = 0 for y in range(__A ): for x in range(__A ): self._process_tile( __A , __A , __A , __A , __A , __A , __A , prompt=__A , num_inference_steps=__A , guidance_scale=__A , noise_level=__A , negative_prompt=__A , num_images_per_prompt=__A , eta=__A , generator=__A , latents=__A , ) current_count += 1 if callback is not None: callback({"progress": current_count / total_tile_count, "image": final_image} ) return final_image def lowercase_( ): '''simple docstring''' lowerCamelCase : Dict = "stabilityai/stable-diffusion-x4-upscaler" lowerCamelCase : Union[str, Any] = StableDiffusionTiledUpscalePipeline.from_pretrained(SCREAMING_SNAKE_CASE_ , revision="fp16" , torch_dtype=torch.floataa ) lowerCamelCase : Optional[Any] = pipe.to("cuda" ) lowerCamelCase : List[str] = Image.open("../../docs/source/imgs/diffusers_library.jpg" ) def callback(SCREAMING_SNAKE_CASE_ ): print(f"""progress: {obj['progress']:.4f}""" ) obj["image"].save("diffusers_library_progress.jpg" ) lowerCamelCase : int = pipe(image=SCREAMING_SNAKE_CASE_ , prompt="Black font, white background, vector" , noise_level=40 , callback=SCREAMING_SNAKE_CASE_ ) final_image.save("diffusers_library.jpg" ) if __name__ == "__main__": main()
283
1
'''simple docstring''' import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def a_ ( __snake_case : Optional[int] ) -> List[str]: """simple docstring""" lowerCamelCase_ =[ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(__snake_case , __snake_case ) def a_ ( __snake_case : List[Any] ) -> int: """simple docstring""" lowerCamelCase_, lowerCamelCase_ =emb.weight.shape lowerCamelCase_ =nn.Linear(__snake_case , __snake_case , bias=__snake_case ) lowerCamelCase_ =emb.weight.data return lin_layer def a_ ( __snake_case : Union[str, Any] , __snake_case : Tuple=None ) -> Dict: """simple docstring""" lowerCamelCase_ ={} for old_key in state_dict.keys(): lowerCamelCase_ =old_key if "moe_layer.experts." in key: if expert_idx is not None: lowerCamelCase_ =key.replace('''moe_layer.experts.0''' , F'''ffn.experts.expert_{expert_idx}''' ) else: lowerCamelCase_ =key.replace('''moe_layer.experts.''' , '''ffn.experts.expert_''' ) if "gate" in key: lowerCamelCase_ =key.replace('''.moe_layer.gate.wg''' , '''.ffn.router.classifier''' ) if "fc2" and "experts" not in key: lowerCamelCase_ =key.replace('''.fc2.''' , '''.ffn.fc2.''' ) if "fc1" and "experts" not in key: lowerCamelCase_ =key.replace('''.fc1.''' , '''.ffn.fc1.''' ) if ".encoder_attn." in key: lowerCamelCase_ =key.replace('''.encoder_attn.''' , '''.cross_attention.''' ) if "encoder_attn_layer_norm" in key: lowerCamelCase_ =key.replace('''encoder_attn_layer_norm''' , '''cross_attention_layer_norm''' ) if "final_layer_norm" in key: lowerCamelCase_ =key.replace('''final_layer_norm''' , '''ff_layer_norm''' ) lowerCamelCase_ =state_dict[old_key] return new_dict def a_ ( __snake_case : Optional[Any] , __snake_case : Any , __snake_case : Optional[int] , __snake_case : Tuple , __snake_case : str = WEIGHTS_NAME ) -> Dict: """simple docstring""" lowerCamelCase_ =[] lowerCamelCase_ =0 os.makedirs(__snake_case , exist_ok=__snake_case ) for expert in range(__snake_case ): lowerCamelCase_ =switch_checkpoint_path + F'''-rank-{expert}.pt''' if os.path.isfile(__snake_case ): lowerCamelCase_ =torch.load(__snake_case )['''model'''] remove_ignore_keys_(__snake_case ) lowerCamelCase_ =rename_fairseq_keys(__snake_case , __snake_case ) lowerCamelCase_ =os.path.join( __snake_case , weights_name.replace('''.bin''' , F'''-{len(__snake_case )+1:05d}-of-???.bin''' ) ) torch.save(__snake_case , __snake_case ) sharded_state_dicts.append(expert_state.keys() ) total_size += sum([value.numel() for key, value in expert_state.items()] ) * dtype_byte_size( expert_state[list(__snake_case )[0]].dtype ) # Add the last block lowerCamelCase_ =os.path.join(__snake_case , weights_name.replace('''.bin''' , F'''-{len(__snake_case )+1:05d}-of-???.bin''' ) ) lowerCamelCase_ =torch.load(switch_checkpoint_path + '''-shared.pt''' )['''model'''] remove_ignore_keys_(__snake_case ) lowerCamelCase_ =rename_fairseq_keys(__snake_case , __snake_case ) lowerCamelCase_ =shared_weights['''decoder.embed_tokens.weight'''] sharded_state_dicts.append(shared_weights.keys() ) # If we only have the shared weights (dummy model/experts saved on the same file) if len(__snake_case ) == 1: lowerCamelCase_ =os.path.join(__snake_case , __snake_case ) torch.save(__snake_case , __snake_case ) return {weights_name: sharded_state_dicts[0]}, None else: torch.save(__snake_case , __snake_case ) # Otherwise, let's build the index lowerCamelCase_ ={} for idx, shard in enumerate(__snake_case ): lowerCamelCase_ =weights_name.replace('''.bin''' , F'''-{idx+1:05d}-of-{len(__snake_case ):05d}.bin''' ) lowerCamelCase_ =os.path.join(__snake_case , weights_name.replace('''.bin''' , F'''-{idx+1:05d}-of-???.bin''' ) ) os.rename(__snake_case , os.path.join(__snake_case , __snake_case ) ) for key in shard: lowerCamelCase_ =shard_file # Add the metadata lowerCamelCase_ ={'''total_size''': total_size} lowerCamelCase_ ={'''metadata''': metadata, '''weight_map''': weight_map} with open(os.path.join(__snake_case , __snake_case ) , '''w''' , encoding='''utf-8''' ) as f: lowerCamelCase_ =json.dumps(__snake_case , indent=2 , sort_keys=__snake_case ) + '''\n''' f.write(__snake_case ) return metadata, index if __name__ == "__main__": a_ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--nllb_moe_checkpoint_path""", default="""/home/arthur_huggingface_co/fairseq/weights/checkpoints/model_moe_54b/checkpoint_2_300000""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--dtype""", default="""float32""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/home/arthur_huggingface_co/fairseq/weights/checkpoints/hf-converted-moe-54b""", type=str, required=False, help="""Path to the output pytorch model.""", ) a_ : Tuple = parser.parse_args() a_ , a_ : int = shard_on_the_fly( args.nllb_moe_checkpoint_path, args.pytorch_dump_folder_path, 1_28, args.dtype, ) a_ : Tuple = NllbMoeConfig.from_pretrained( """facebook/nllb-200-3.3B""", encoder_sparse_step=4, decoder_sparse_step=4, num_experts=1_28 ) config.save_pretrained(args.pytorch_dump_folder_path) a_ : Any = NllbMoeModel.from_pretrained(args.pytorch_dump_folder_path) print("""Done""") model.save_pretrained(args.pytorch_dump_folder_path)
6
'''simple docstring''' a_ : List[Any] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def a_ ( __snake_case : int ) -> int: """simple docstring""" lowerCamelCase_ =0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution a_ : list[bool | None] = [None] * 10_00_00_00 a_ : List[Any] = True a_ : Optional[Any] = False def a_ ( __snake_case : int ) -> bool: """simple docstring""" if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore lowerCamelCase_ =chain(next_number(__snake_case ) ) lowerCamelCase_ =number_chain while number < 1000_0000: lowerCamelCase_ =number_chain number *= 10 return number_chain def a_ ( __snake_case : int = 1000_0000 ) -> int: """simple docstring""" for i in range(1 , __snake_case ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__snake_case ) if __name__ == "__main__": import doctest doctest.testmod() print(F"""{solution() = }""")
6
1
from __future__ import annotations def lowerCamelCase__ ( A__ : list[int] , A__ : list[int] , A__ : int ): '''simple docstring''' __lowerCamelCase = list(range(len(A__ ) ) ) __lowerCamelCase = [v / w for v, w in zip(A__ , A__ )] index.sort(key=lambda A__ : ratio[i] , reverse=A__ ) __lowerCamelCase = 0 __lowerCamelCase = [0] * len(A__ ) for i in index: if weight[i] <= capacity: __lowerCamelCase = 1 max_value += value[i] capacity -= weight[i] else: __lowerCamelCase = capacity / weight[i] max_value += value[i] * capacity / weight[i] break return max_value, fractions if __name__ == "__main__": import doctest doctest.testmod()
12
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging UpperCAmelCase_ = logging.get_logger(__name__) class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : Tuple = ['pixel_values'] def __init__( self: Any , UpperCamelCase_: bool = True , UpperCamelCase_: Union[int, float] = 1 / 2_55 , UpperCamelCase_: bool = True , UpperCamelCase_: int = 8 , **UpperCamelCase_: Tuple , ): super().__init__(**UpperCamelCase_ ) __lowerCamelCase = do_rescale __lowerCamelCase = rescale_factor __lowerCamelCase = do_pad __lowerCamelCase = pad_size def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: np.ndarray , UpperCamelCase_: float , UpperCamelCase_: Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase_: Tuple ): return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: Union[str, Any] , UpperCamelCase_: np.ndarray , UpperCamelCase_: int , UpperCamelCase_: Optional[Union[str, ChannelDimension]] = None ): __lowerCamelCase, __lowerCamelCase = get_image_size(UpperCamelCase_ ) __lowerCamelCase = (old_height // size + 1) * size - old_height __lowerCamelCase = (old_width // size + 1) * size - old_width return pad(UpperCamelCase_ , ((0, pad_height), (0, pad_width)) , mode="""symmetric""" , data_format=UpperCamelCase_ ) def lowerCAmelCase__ ( self: str , UpperCamelCase_: ImageInput , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[float] = None , UpperCamelCase_: Optional[bool] = None , UpperCamelCase_: Optional[int] = None , UpperCamelCase_: Optional[Union[str, TensorType]] = None , UpperCamelCase_: Union[str, ChannelDimension] = ChannelDimension.FIRST , **UpperCamelCase_: Any , ): __lowerCamelCase = do_rescale if do_rescale is not None else self.do_rescale __lowerCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowerCamelCase = do_pad if do_pad is not None else self.do_pad __lowerCamelCase = pad_size if pad_size is not None else self.pad_size __lowerCamelCase = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) # All transformations expect numpy arrays. __lowerCamelCase = [to_numpy_array(UpperCamelCase_ ) for image in images] if do_rescale: __lowerCamelCase = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_pad: __lowerCamelCase = [self.pad(UpperCamelCase_ , size=UpperCamelCase_ ) for image in images] __lowerCamelCase = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] __lowerCamelCase = {"""pixel_values""": images} return BatchFeature(data=UpperCamelCase_ , tensor_type=UpperCamelCase_ )
12
1
"""simple docstring""" import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a : Union[str, Any]= logging.get_logger(__name__) _a : List[str]= { "vocab_file": "vocab.json", "tokenizer_config_file": "tokenizer_config.json", "merges_file": "merges.txt", } _a : List[str]= { "vocab_file": { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json" ), }, "tokenizer_config_file": { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json" ), }, "merges_file": { "facebook/s2t-wav2vec2-large-en-de": ( "https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt" ), }, } _a : List[str]= "</w>" _a : Dict= "@@ " def __UpperCAmelCase ( UpperCAmelCase_ : int ) -> Union[str, Any]: '''simple docstring''' __snake_case : Optional[Any] = set() __snake_case : Optional[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __snake_case : List[str] = char return pairs # Speech2Text2 has no max input length _a : Union[str, Any]= {"facebook/s2t-wav2vec2-large-en-de": 1_024} class UpperCamelCase ( lowercase ): UpperCAmelCase : List[str] = VOCAB_FILES_NAMES UpperCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : int = ["""input_ids""", """attention_mask"""] def __init__(self : str , _A : Any , _A : Union[str, Any]="<s>" , _A : Optional[int]="<pad>" , _A : Tuple="</s>" , _A : List[Any]="<unk>" , _A : List[str]=False , _A : List[Any]=None , **_A : int , ) -> List[str]: super().__init__( unk_token=_A , bos_token=_A , eos_token=_A , pad_token=_A , do_lower_case=_A , **_A , ) __snake_case : List[str] = do_lower_case with open(_A , encoding='utf-8') as vocab_handle: __snake_case : Optional[Any] = json.load(_A) __snake_case : Dict = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"No merges files provided. {self.__class__.__name__} can only be used for decoding.") __snake_case : List[Any] = None __snake_case : List[str] = None else: with open(_A , encoding='utf-8') as merges_handle: __snake_case : Tuple = merges_handle.read().split('\n')[:-1] __snake_case : List[Any] = [tuple(merge.split()[:2]) for merge in merges] __snake_case : List[str] = dict(zip(_A , range(len(_A)))) __snake_case : Tuple = {} @property def _lowercase (self : Dict) -> int: return len(self.decoder) def _lowercase (self : Tuple) -> Dict: return dict(self.encoder , **self.added_tokens_encoder) def _lowercase (self : Optional[int] , _A : Dict) -> Any: __snake_case : Tuple = tuple(token[:-1]) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] __snake_case : List[Any] = get_pairs(_A) if not pairs: return token while True: __snake_case : Dict = min(_A , key=lambda _A: self.bpe_ranks.get(_A , float('inf'))) if bigram not in self.bpe_ranks: break __snake_case : Dict = bigram __snake_case : Union[str, Any] = [] __snake_case : List[Any] = 0 while i < len(_A): try: __snake_case : Any = word.index(_A , _A) except ValueError: new_word.extend(word[i:]) break else: new_word.extend(word[i:j]) __snake_case : List[Any] = j if word[i] == first and i < len(_A) - 1 and word[i + 1] == second: new_word.append(first + second) i += 2 else: new_word.append(word[i]) i += 1 __snake_case : Tuple = tuple(_A) __snake_case : List[str] = new_word if len(_A) == 1: break else: __snake_case : int = get_pairs(_A) __snake_case : Tuple = ' '.join(_A) if word == "\n " + BPE_TOKEN_MERGES: __snake_case : Tuple = '\n' + BPE_TOKEN_MERGES if word.endswith(_A): __snake_case : List[str] = word.replace(_A , '') __snake_case : Tuple = word.replace(' ' , _A) __snake_case : Optional[int] = word return word def _lowercase (self : str , _A : List[str]) -> str: if self.bpe_ranks is None: raise ValueError( 'This tokenizer was instantiated without a `merges.txt` file, so' ' that it can only be used for decoding, not for encoding.' 'Make sure to provide `merges.txt` file at instantiation to enable ' 'encoding.') if self.do_lower_case: __snake_case : Dict = text.lower() __snake_case : str = text.split() __snake_case : Optional[int] = [] for token in text: if token: split_tokens.extend(list(self.bpe(_A).split(' '))) return split_tokens def _lowercase (self : str , _A : str) -> int: return self.encoder.get(_A , self.encoder.get(self.unk_token)) def _lowercase (self : int , _A : int) -> str: __snake_case : str = self.decoder.get(_A , self.unk_token) return result def _lowercase (self : Optional[int] , _A : List[str]) -> str: __snake_case : Union[str, Any] = ' '.join(_A) # make sure @@ tokens are concatenated __snake_case : Tuple = ''.join(string.split(_A)) return string def _lowercase (self : str , _A : str , _A : Optional[str] = None) -> Tuple[str]: if not os.path.isdir(_A): logger.error(f"Vocabulary path ({save_directory}) should be a directory") return __snake_case : Any = os.path.join( _A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) __snake_case : List[Any] = os.path.join( _A , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file']) with open(_A , 'w' , encoding='utf-8') as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_A , ensure_ascii=_A) + '\n') __snake_case : int = 0 if self.bpe_ranks is None: return (vocab_file,) with open(_A , 'w' , encoding='utf-8') as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _A: kv[1]): if index != token_index: logger.warning( f"Saving vocabulary to {merges_file}: BPE merge indices are not consecutive." ' Please check that the tokenizer is not corrupted!') __snake_case : str = token_index writer.write(' '.join(_A) + '\n') index += 1 return (vocab_file, merges_file)
356
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer _a : Any= logging.get_logger(__name__) _a : str= {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} _a : Optional[Any]= [ "small", "small-base", "medium", "medium-base", "intermediate", "intermediate-base", "large", "large-base", "xlarge", "xlarge-base", ] _a : List[Any]= { "vocab_file": { "funnel-transformer/small": "https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt", "funnel-transformer/small-base": "https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt", "funnel-transformer/medium": "https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt", "funnel-transformer/medium-base": ( "https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt" ), "funnel-transformer/intermediate": ( "https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt" ), "funnel-transformer/intermediate-base": ( "https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt" ), "funnel-transformer/large": "https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt", "funnel-transformer/large-base": "https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt", "funnel-transformer/xlarge": "https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt", "funnel-transformer/xlarge-base": ( "https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt" ), }, "tokenizer_file": { "funnel-transformer/small": "https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json", "funnel-transformer/small-base": ( "https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json" ), "funnel-transformer/medium": "https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json", "funnel-transformer/medium-base": ( "https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json" ), "funnel-transformer/intermediate": ( "https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json" ), "funnel-transformer/intermediate-base": ( "https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json" ), "funnel-transformer/large": "https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json", "funnel-transformer/large-base": ( "https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json" ), "funnel-transformer/xlarge": "https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json", "funnel-transformer/xlarge-base": ( "https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json" ), }, } _a : str= {f'''funnel-transformer/{name}''': 512 for name in _model_names} _a : List[Any]= {f'''funnel-transformer/{name}''': {"do_lower_case": True} for name in _model_names} class UpperCamelCase ( lowercase ): UpperCAmelCase : List[str] = VOCAB_FILES_NAMES UpperCAmelCase : Tuple = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : int = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase : Tuple = FunnelTokenizer UpperCAmelCase : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : int = 2 def __init__(self : int , _A : Any=None , _A : Union[str, Any]=None , _A : Union[str, Any]=True , _A : List[str]="<unk>" , _A : Any="<sep>" , _A : Dict="<pad>" , _A : Tuple="<cls>" , _A : Dict="<mask>" , _A : Optional[Any]="<s>" , _A : List[Any]="</s>" , _A : Optional[int]=True , _A : Dict=True , _A : Tuple=None , _A : int="##" , **_A : Any , ) -> str: super().__init__( _A , tokenizer_file=_A , do_lower_case=_A , unk_token=_A , sep_token=_A , pad_token=_A , cls_token=_A , mask_token=_A , bos_token=_A , eos_token=_A , clean_text=_A , tokenize_chinese_chars=_A , strip_accents=_A , wordpieces_prefix=_A , **_A , ) __snake_case : Optional[int] = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get('lowercase' , _A) != do_lower_case or normalizer_state.get('strip_accents' , _A) != strip_accents or normalizer_state.get('handle_chinese_chars' , _A) != tokenize_chinese_chars ): __snake_case : List[str] = getattr(_A , normalizer_state.pop('type')) __snake_case : int = do_lower_case __snake_case : Optional[int] = strip_accents __snake_case : str = tokenize_chinese_chars __snake_case : Optional[int] = normalizer_class(**_A) __snake_case : str = do_lower_case def _lowercase (self : Optional[Any] , _A : Dict , _A : Tuple=None) -> Any: __snake_case : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase (self : str , _A : List[int] , _A : Optional[List[int]] = None) -> List[int]: __snake_case : Union[str, Any] = [self.sep_token_id] __snake_case : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls) * [self.cls_token_type_id] + len(token_ids_a + sep) * [0] return len(cls) * [self.cls_token_type_id] + len(token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _lowercase (self : Tuple , _A : str , _A : Optional[str] = None) -> Tuple[str]: __snake_case : int = self._tokenizer.model.save(_A , name=_A) return tuple(_A)
95
0
"""simple docstring""" from __future__ import annotations def lowercase (snake_case__ : list[float] , snake_case__ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' print(f'''Vertex\tShortest Distance from vertex {src}''' ) for i, d in enumerate(snake_case__ ): print(f'''{i}\t\t{d}''' ) def lowercase (snake_case__ : list[dict[str, int]] , snake_case__ : list[float] , snake_case__ : int ) -> str: '''simple docstring''' for j in range(snake_case__ ): lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: return True return False def lowercase (snake_case__ : list[dict[str, int]] , snake_case__ : int , snake_case__ : int , snake_case__ : int ) -> list[float]: '''simple docstring''' lowerCAmelCase = [float("""inf""" )] * vertex_count lowerCAmelCase = 0.0 for _ in range(vertex_count - 1 ): for j in range(snake_case__ ): lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = (graph[j][k] for k in ["""src""", """dst""", """weight"""]) if distance[u] != float("""inf""" ) and distance[u] + w < distance[v]: lowerCAmelCase = distance[u] + w lowerCAmelCase = check_negative_cycle(snake_case__ , snake_case__ , snake_case__ ) if negative_cycle_exists: raise Exception("""Negative cycle found""" ) return distance if __name__ == "__main__": import doctest doctest.testmod() a = int(input('Enter number of vertices: ').strip()) a = int(input('Enter number of edges: ').strip()) a = [{} for _ in range(E)] for i in range(E): print('Edge ', i + 1) a , a , a = ( int(x) for x in input('Enter source, destination, weight: ').strip().split(' ') ) a = {'src': src, 'dst': dest, 'weight': weight} a = int(input('\nEnter shortest path source:').strip()) a = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
155
"""simple docstring""" from itertools import permutations def lowercase (snake_case__ : tuple ) -> bool: '''simple docstring''' if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase = [7, 11, 13, 17] for i, test in enumerate(snake_case__ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowercase (snake_case__ : int = 10 ) -> int: '''simple docstring''' return sum( int("""""".join(map(snake_case__ , snake_case__ ) ) ) for num in permutations(range(snake_case__ ) ) if is_substring_divisible(snake_case__ ) ) if __name__ == "__main__": print(f"""{solution() = }""")
155
1
"""simple docstring""" from math import isqrt, loga def __snake_case ( SCREAMING_SNAKE_CASE__ : int ) -> list[int]: '''simple docstring''' _UpperCAmelCase : Union[str, Any] = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): _UpperCAmelCase : List[str] = False return [i for i in range(2 , SCREAMING_SNAKE_CASE__ ) if is_prime[i]] def __snake_case ( SCREAMING_SNAKE_CASE__ : int = 800_800 , SCREAMING_SNAKE_CASE__ : int = 800_800 ) -> int: '''simple docstring''' _UpperCAmelCase : Tuple = degree * loga(SCREAMING_SNAKE_CASE__ ) _UpperCAmelCase : List[str] = int(SCREAMING_SNAKE_CASE__ ) _UpperCAmelCase : Any = calculate_prime_numbers(SCREAMING_SNAKE_CASE__ ) _UpperCAmelCase : Optional[Any] = 0 _UpperCAmelCase : Union[str, Any] = 0 _UpperCAmelCase : Optional[Any] = len(SCREAMING_SNAKE_CASE__ ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(F"{solution() = }")
360
"""simple docstring""" import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _lowerCAmelCase : Optional[int] = logging.get_logger(__name__) _lowerCAmelCase : Dict = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} _lowerCAmelCase : Tuple = { "vocab_file": { "allenai/longformer-base-4096": "https://huggingface.co/allenai/longformer-base-4096/resolve/main/vocab.json", "allenai/longformer-large-4096": ( "https://huggingface.co/allenai/longformer-large-4096/resolve/main/vocab.json" ), "allenai/longformer-large-4096-finetuned-triviaqa": ( "https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/vocab.json" ), "allenai/longformer-base-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/vocab.json" ), "allenai/longformer-large-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/vocab.json" ), }, "merges_file": { "allenai/longformer-base-4096": "https://huggingface.co/allenai/longformer-base-4096/resolve/main/merges.txt", "allenai/longformer-large-4096": ( "https://huggingface.co/allenai/longformer-large-4096/resolve/main/merges.txt" ), "allenai/longformer-large-4096-finetuned-triviaqa": ( "https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/merges.txt" ), "allenai/longformer-base-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/merges.txt" ), "allenai/longformer-large-4096-extra.pos.embd.only": ( "https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/merges.txt" ), }, } _lowerCAmelCase : str = { "allenai/longformer-base-4096": 40_96, "allenai/longformer-large-4096": 40_96, "allenai/longformer-large-4096-finetuned-triviaqa": 40_96, "allenai/longformer-base-4096-extra.pos.embd.only": 40_96, "allenai/longformer-large-4096-extra.pos.embd.only": 40_96, } @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def __snake_case ( ) -> List[Any]: '''simple docstring''' _UpperCAmelCase : str = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) _UpperCAmelCase : Any = bs[:] _UpperCAmelCase : Tuple = 0 for b in range(2**8 ): if b not in bs: bs.append(SCREAMING_SNAKE_CASE__ ) cs.append(2**8 + n ) n += 1 _UpperCAmelCase : Union[str, Any] = [chr(SCREAMING_SNAKE_CASE__ ) for n in cs] return dict(zip(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) def __snake_case ( SCREAMING_SNAKE_CASE__ : Optional[Any] ) -> int: '''simple docstring''' _UpperCAmelCase : int = set() _UpperCAmelCase : int = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _UpperCAmelCase : Optional[int] = char return pairs class UpperCAmelCase_ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE : List[str] = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE : List[str] = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self : Optional[Any] , A : int , A : Any , A : List[str]="replace" , A : List[Any]="<s>" , A : int="</s>" , A : Union[str, Any]="</s>" , A : Tuple="<s>" , A : str="<unk>" , A : Dict="<pad>" , A : Optional[Any]="<mask>" , A : Tuple=False , **A : Dict , ): _UpperCAmelCase : Tuple = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else bos_token _UpperCAmelCase : int = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else eos_token _UpperCAmelCase : int = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else sep_token _UpperCAmelCase : Dict = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else cls_token _UpperCAmelCase : Optional[int] = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else unk_token _UpperCAmelCase : Dict = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _UpperCAmelCase : Dict = AddedToken(A , lstrip=A , rstrip=A ) if isinstance(A , A ) else mask_token super().__init__( errors=A , bos_token=A , eos_token=A , unk_token=A , sep_token=A , cls_token=A , pad_token=A , mask_token=A , add_prefix_space=A , **A , ) with open(A , encoding="utf-8" ) as vocab_handle: _UpperCAmelCase : Union[str, Any] = json.load(A ) _UpperCAmelCase : List[str] = {v: k for k, v in self.encoder.items()} _UpperCAmelCase : Dict = errors # how to handle errors in decoding _UpperCAmelCase : List[str] = bytes_to_unicode() _UpperCAmelCase : Dict = {v: k for k, v in self.byte_encoder.items()} with open(A , encoding="utf-8" ) as merges_handle: _UpperCAmelCase : Union[str, Any] = merges_handle.read().split("\n" )[1:-1] _UpperCAmelCase : Union[str, Any] = [tuple(merge.split() ) for merge in bpe_merges] _UpperCAmelCase : List[str] = dict(zip(A , range(len(A ) ) ) ) _UpperCAmelCase : Tuple = {} _UpperCAmelCase : Optional[Any] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _UpperCAmelCase : Optional[int] = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property def snake_case_ ( self : Optional[Any] ): return len(self.encoder ) def snake_case_ ( self : List[str] ): return dict(self.encoder , **self.added_tokens_encoder ) def snake_case_ ( self : Tuple , A : Union[str, Any] ): if token in self.cache: return self.cache[token] _UpperCAmelCase : Optional[int] = tuple(A ) _UpperCAmelCase : Optional[Any] = get_pairs(A ) if not pairs: return token while True: _UpperCAmelCase : Optional[int] = min(A , key=lambda A : self.bpe_ranks.get(A , float("inf" ) ) ) if bigram not in self.bpe_ranks: break _UpperCAmelCase , _UpperCAmelCase : str = bigram _UpperCAmelCase : Dict = [] _UpperCAmelCase : Union[str, Any] = 0 while i < len(A ): try: _UpperCAmelCase : int = word.index(A , A ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _UpperCAmelCase : Dict = j if word[i] == first and i < len(A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _UpperCAmelCase : Optional[Any] = tuple(A ) _UpperCAmelCase : Dict = new_word if len(A ) == 1: break else: _UpperCAmelCase : Optional[int] = get_pairs(A ) _UpperCAmelCase : Any = " ".join(A ) _UpperCAmelCase : int = word return word def snake_case_ ( self : Optional[int] , A : List[str] ): _UpperCAmelCase : str = [] for token in re.findall(self.pat , A ): _UpperCAmelCase : Optional[Any] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(A ).split(" " ) ) return bpe_tokens def snake_case_ ( self : Optional[int] , A : Union[str, Any] ): return self.encoder.get(A , self.encoder.get(self.unk_token ) ) def snake_case_ ( self : Union[str, Any] , A : List[str] ): return self.decoder.get(A ) def snake_case_ ( self : Dict , A : int ): _UpperCAmelCase : Tuple = "".join(A ) _UpperCAmelCase : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def snake_case_ ( self : int , A : str , A : Optional[str] = None ): if not os.path.isdir(A ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return _UpperCAmelCase : List[Any] = os.path.join( A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) _UpperCAmelCase : Any = os.path.join( A , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(A , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=A , ensure_ascii=A ) + "\n" ) _UpperCAmelCase : Optional[Any] = 0 with open(A , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda A : kv[1] ): if index != token_index: logger.warning( f'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' " Please check that the tokenizer is not corrupted!" ) _UpperCAmelCase : Any = token_index writer.write(" ".join(A ) + "\n" ) index += 1 return vocab_file, merge_file def snake_case_ ( self : int , A : List[int] , A : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _UpperCAmelCase : List[Any] = [self.cls_token_id] _UpperCAmelCase : List[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def snake_case_ ( self : Optional[Any] , A : List[int] , A : Optional[List[int]] = None , A : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A , token_ids_a=A , already_has_special_tokens=A ) if token_ids_a is None: return [1] + ([0] * len(A )) + [1] return [1] + ([0] * len(A )) + [1, 1] + ([0] * len(A )) + [1] def snake_case_ ( self : str , A : List[int] , A : Optional[List[int]] = None ): _UpperCAmelCase : Optional[Any] = [self.sep_token_id] _UpperCAmelCase : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case_ ( self : str , A : List[str] , A : List[str]=False , **A : List[str] ): _UpperCAmelCase : Optional[Any] = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(A ) > 0 and not text[0].isspace()): _UpperCAmelCase : List[Any] = " " + text return (text, kwargs)
202
0
"""simple docstring""" lowercase__ : Union[str, Any] = [ [0, 1_6, 1_3, 0, 0, 0], [0, 0, 1_0, 1_2, 0, 0], [0, 4, 0, 0, 1_4, 0], [0, 0, 9, 0, 0, 2_0], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def UpperCamelCase_ ( lowerCAmelCase__ : int , lowerCAmelCase__ : Any , lowerCAmelCase__ : str , lowerCAmelCase__ : Dict ) -> Dict: """simple docstring""" lowerCAmelCase_ : str = [False] * len(snake_case_ ) lowerCAmelCase_ : int = [s] lowerCAmelCase_ : List[str] = True while queue: lowerCAmelCase_ : List[Any] = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(snake_case_ ) lowerCAmelCase_ : Tuple = True lowerCAmelCase_ : Optional[Any] = u return visited[t] def UpperCamelCase_ ( lowerCAmelCase__ : Any , lowerCAmelCase__ : str , lowerCAmelCase__ : List[Any] ) -> Optional[int]: """simple docstring""" lowerCAmelCase_ : Optional[int] = [-1] * (len(snake_case_ )) lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : str = [i[:] for i in graph] # Record original cut, copy. while bfs(snake_case_ , snake_case_ , snake_case_ , snake_case_ ): lowerCAmelCase_ : Optional[Any] = float('Inf' ) lowerCAmelCase_ : List[Any] = sink while s != source: # Find the minimum value in select path lowerCAmelCase_ : int = min(snake_case_ , graph[parent[s]][s] ) lowerCAmelCase_ : Dict = parent[s] max_flow += path_flow lowerCAmelCase_ : Union[str, Any] = sink while v != source: lowerCAmelCase_ : Tuple = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow lowerCAmelCase_ : List[Any] = parent[v] for i in range(len(snake_case_ ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
224
"""simple docstring""" from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
126
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxSeqaSeqConfigWithPast from ...utils import logging if TYPE_CHECKING: from ...feature_extraction_utils import FeatureExtractionMixin from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType __snake_case = logging.get_logger(__name__) __snake_case = { '''openai/whisper-base''': '''https://huggingface.co/openai/whisper-base/resolve/main/config.json''', } # fmt: off __snake_case = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 3_57, 3_66, 4_38, 5_32, 6_85, 7_05, 7_96, 9_30, 10_58, 12_20, 12_67, 12_79, 13_03, 13_43, 13_77, 13_91, 16_35, 17_82, 18_75, 21_62, 23_61, 24_88, 34_67, 40_08, 42_11, 46_00, 48_08, 52_99, 58_55, 63_29, 72_03, 96_09, 99_59, 1_05_63, 1_07_86, 1_14_20, 1_17_09, 1_19_07, 1_31_63, 1_36_97, 1_37_00, 1_48_08, 1_53_06, 1_64_10, 1_67_91, 1_79_92, 1_92_03, 1_95_10, 2_07_24, 2_23_05, 2_29_35, 2_70_07, 3_01_09, 3_04_20, 3_34_09, 3_49_49, 4_02_83, 4_04_93, 4_05_49, 4_72_82, 4_91_46, 5_02_57, 5_03_59, 5_03_60, 5_03_61 ] __snake_case = [ 1, 2, 7, 8, 9, 10, 14, 25, 26, 27, 28, 29, 31, 58, 59, 60, 61, 62, 63, 90, 91, 92, 93, 3_59, 5_03, 5_22, 5_42, 8_73, 8_93, 9_02, 9_18, 9_22, 9_31, 13_50, 18_53, 19_82, 24_60, 26_27, 32_46, 32_53, 32_68, 35_36, 38_46, 39_61, 41_83, 46_67, 65_85, 66_47, 72_73, 90_61, 93_83, 1_04_28, 1_09_29, 1_19_38, 1_20_33, 1_23_31, 1_25_62, 1_37_93, 1_41_57, 1_46_35, 1_52_65, 1_56_18, 1_65_53, 1_66_04, 1_83_62, 1_89_56, 2_00_75, 2_16_75, 2_25_20, 2_61_30, 2_61_61, 2_64_35, 2_82_79, 2_94_64, 3_16_50, 3_23_02, 3_24_70, 3_68_65, 4_28_63, 4_74_25, 4_98_70, 5_02_54, 5_02_58, 5_03_60, 5_03_61, 5_03_62 ] class __snake_case ( lowerCamelCase__ ): __lowerCamelCase : List[Any] = """whisper""" __lowerCamelCase : List[str] = ["""past_key_values"""] __lowerCamelCase : List[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , snake_case__=5_1865 , snake_case__=80 , snake_case__=6 , snake_case__=4 , snake_case__=6 , snake_case__=4 , snake_case__=1536 , snake_case__=1536 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=5_0257 , snake_case__=True , snake_case__=True , snake_case__="gelu" , snake_case__=256 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.0 , snake_case__=0.02 , snake_case__=False , snake_case__=1500 , snake_case__=448 , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=5_0256 , snake_case__=None , snake_case__=[220, 5_0256] , snake_case__=False , snake_case__=256 , snake_case__=False , snake_case__=0.05 , snake_case__=10 , snake_case__=2 , snake_case__=0.0 , snake_case__=10 , snake_case__=0 , snake_case__=7 , **snake_case__ , ) -> Dict: '''simple docstring''' UpperCAmelCase : Union[str, Any] =vocab_size UpperCAmelCase : Tuple =num_mel_bins UpperCAmelCase : Optional[int] =d_model UpperCAmelCase : str =encoder_layers UpperCAmelCase : str =encoder_attention_heads UpperCAmelCase : Optional[Any] =decoder_layers UpperCAmelCase : Optional[int] =decoder_attention_heads UpperCAmelCase : List[Any] =decoder_ffn_dim UpperCAmelCase : int =encoder_ffn_dim UpperCAmelCase : Dict =dropout UpperCAmelCase : Any =attention_dropout UpperCAmelCase : List[str] =activation_dropout UpperCAmelCase : List[str] =activation_function UpperCAmelCase : Optional[Any] =init_std UpperCAmelCase : Any =encoder_layerdrop UpperCAmelCase : int =decoder_layerdrop UpperCAmelCase : Dict =use_cache UpperCAmelCase : Union[str, Any] =encoder_layers UpperCAmelCase : Optional[Any] =scale_embedding # scale factor will be sqrt(d_model) if True UpperCAmelCase : List[Any] =max_source_positions UpperCAmelCase : List[str] =max_target_positions # Audio Classification-specific parameters. Feel free to ignore for other classes. UpperCAmelCase : Any =classifier_proj_size UpperCAmelCase : int =use_weighted_layer_sum # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase : int =apply_spec_augment UpperCAmelCase : List[Any] =mask_time_prob UpperCAmelCase : Optional[Any] =mask_time_length UpperCAmelCase : Optional[int] =mask_time_min_masks UpperCAmelCase : Union[str, Any] =mask_feature_prob UpperCAmelCase : Optional[int] =mask_feature_length UpperCAmelCase : Union[str, Any] =mask_feature_min_masks UpperCAmelCase : Any =median_filter_width super().__init__( pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ , is_encoder_decoder=snake_case__ , decoder_start_token_id=snake_case__ , suppress_tokens=snake_case__ , begin_suppress_tokens=snake_case__ , **snake_case__ , ) class __snake_case ( lowerCamelCase__ ): @property def UpperCAmelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' UpperCAmelCase : List[str] =OrderedDict( [ ('''input_features''', {0: '''batch''', 1: '''feature_size''', 2: '''encoder_sequence'''}), ] ) if self.use_past: UpperCAmelCase : Tuple ={0: '''batch'''} else: UpperCAmelCase : Optional[Any] ={0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(snake_case__ , direction='''inputs''' ) return common_inputs def UpperCAmelCase__ ( self , snake_case__ , snake_case__ = -1 , snake_case__ = -1 , snake_case__ = False , snake_case__ = None , snake_case__ = 2_2050 , snake_case__ = 5.0 , snake_case__ = 220 , ) -> Mapping[str, Any]: '''simple docstring''' UpperCAmelCase : List[Any] =OrderedDict() UpperCAmelCase : Tuple =OnnxConfig.generate_dummy_inputs( self , preprocessor=preprocessor.feature_extractor , batch_size=snake_case__ , framework=snake_case__ , sampling_rate=snake_case__ , time_duration=snake_case__ , frequency=snake_case__ , ) UpperCAmelCase : str =encoder_inputs['''input_features'''].shape[2] UpperCAmelCase : str =encoder_sequence_length // 2 if self.use_past else seq_length UpperCAmelCase : Union[str, Any] =super().generate_dummy_inputs( preprocessor.tokenizer , snake_case__ , snake_case__ , snake_case__ , snake_case__ ) UpperCAmelCase : Dict =encoder_inputs.pop('''input_features''' ) UpperCAmelCase : Tuple =decoder_inputs.pop('''decoder_input_ids''' ) if "past_key_values" in decoder_inputs: UpperCAmelCase : List[Any] =decoder_inputs.pop('''past_key_values''' ) return dummy_inputs @property def UpperCAmelCase__ ( self ) -> float: '''simple docstring''' return 1e-3
370
from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def lowerCAmelCase_ ( __lowerCAmelCase = "isbn/0140328726" )-> dict: '''simple docstring''' UpperCAmelCase : Tuple =olid.strip().strip('''/''' ) # Remove leading/trailing whitespace & slashes if new_olid.count('''/''' ) != 1: UpperCAmelCase : Any =f'''{olid} is not a valid Open Library olid''' raise ValueError(__lowerCAmelCase ) return requests.get(f'''https://openlibrary.org/{new_olid}.json''' ).json() def lowerCAmelCase_ ( __lowerCAmelCase )-> dict: '''simple docstring''' UpperCAmelCase : int ={ '''title''': '''Title''', '''publish_date''': '''Publish date''', '''authors''': '''Authors''', '''number_of_pages''': '''Number of pages:''', '''first_sentence''': '''First sentence''', '''isbn_10''': '''ISBN (10)''', '''isbn_13''': '''ISBN (13)''', } UpperCAmelCase : Tuple ={better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCAmelCase : Optional[int] =[ get_openlibrary_data(author['''key'''] )['''name'''] for author in data['''Authors'''] ] UpperCAmelCase : List[str] =data['''First sentence''']['''value'''] for key, value in data.items(): if isinstance(__lowerCAmelCase , __lowerCAmelCase ): UpperCAmelCase : str =''', '''.join(__lowerCAmelCase ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: __snake_case = input('''\nEnter the ISBN code to search (or \'quit\' to stop): ''').strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(f'Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.') continue print(f'\nSearching Open Library for ISBN: {isbn}...\n') try: __snake_case = summarize_book(get_openlibrary_data(f'isbn/{isbn}')) print('''\n'''.join(f'{key}: {value}' for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(f'Sorry, there are no results for ISBN: {isbn}.')
78
0
"""simple docstring""" import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def __a ( ) ->List[Any]: with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(_SCREAMING_SNAKE_CASE ): requests.request('GET' , 'https://huggingface.co' ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request('GET' , 'https://huggingface.co' , timeout=1.0 ) @pytest.mark.integration def __a ( ) ->List[str]: with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request('GET' , 'https://huggingface.co' ) def __a ( ) ->Any: with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(_SCREAMING_SNAKE_CASE ): http_head('https://huggingface.co' )
290
"""simple docstring""" import pytest import datasets # Import fixture modules as plugins __A : int = ["tests.fixtures.files", "tests.fixtures.hub", "tests.fixtures.fsspec"] def lowercase ( _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' for item in items: if any(marker in item.keywords for marker in ['''integration''', '''unit'''] ): continue item.add_marker(pytest.mark.unit ) def lowercase ( _SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' config.addinivalue_line('''markers''' , '''torchaudio_latest: mark test to run with torchaudio>=0.12''' ) @pytest.fixture(autouse=_SCREAMING_SNAKE_CASE ) def lowercase ( _SCREAMING_SNAKE_CASE : Optional[int] , _SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' _UpperCAmelCase = tmp_path_factory.getbasetemp() / '''cache''' _UpperCAmelCase = test_hf_cache_home / '''datasets''' _UpperCAmelCase = test_hf_cache_home / '''metrics''' _UpperCAmelCase = test_hf_cache_home / '''modules''' monkeypatch.setattr('''datasets.config.HF_DATASETS_CACHE''' , str(_SCREAMING_SNAKE_CASE ) ) monkeypatch.setattr('''datasets.config.HF_METRICS_CACHE''' , str(_SCREAMING_SNAKE_CASE ) ) monkeypatch.setattr('''datasets.config.HF_MODULES_CACHE''' , str(_SCREAMING_SNAKE_CASE ) ) _UpperCAmelCase = test_hf_datasets_cache / '''downloads''' monkeypatch.setattr('''datasets.config.DOWNLOADED_DATASETS_PATH''' , str(_SCREAMING_SNAKE_CASE ) ) _UpperCAmelCase = test_hf_datasets_cache / '''downloads''' / '''extracted''' monkeypatch.setattr('''datasets.config.EXTRACTED_DATASETS_PATH''' , str(_SCREAMING_SNAKE_CASE ) ) @pytest.fixture(autouse=_SCREAMING_SNAKE_CASE , scope='''session''' ) def lowercase ( ): '''simple docstring''' datasets.disable_progress_bar() @pytest.fixture(autouse=_SCREAMING_SNAKE_CASE ) def lowercase ( _SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' monkeypatch.setattr('''datasets.config.HF_UPDATE_DOWNLOAD_COUNTS''' , _SCREAMING_SNAKE_CASE ) @pytest.fixture def lowercase ( _SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' monkeypatch.setattr('''sqlalchemy.util.deprecations.SILENCE_UBER_WARNING''' , _SCREAMING_SNAKE_CASE )
260
0
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class UpperCAmelCase__ ( A_ ): """simple docstring""" UpperCAmelCase__ : Tuple = ["image_processor"] UpperCAmelCase__ : Optional[Any] = "SamImageProcessor" def __init__( self , A_ ) -> List[Any]: super().__init__(A_ ) __UpperCamelCase =self.image_processor __UpperCamelCase =-10 __UpperCamelCase =self.image_processor.size['longest_edge'] def __call__( self , A_=None , A_=None , A_=None , A_=None , A_ = None , **A_ , ) -> BatchEncoding: __UpperCamelCase =self.image_processor( A_ , return_tensors=A_ , **A_ , ) # pop arguments that are not used in the foward but used nevertheless __UpperCamelCase =encoding_image_processor['original_sizes'] if hasattr(A_ , 'numpy' ): # Checks if Torch or TF tensor __UpperCamelCase =original_sizes.numpy() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase =self._check_and_preprocess_points( input_points=A_ , input_labels=A_ , input_boxes=A_ , ) __UpperCamelCase =self._normalize_and_convert( A_ , A_ , input_points=A_ , input_labels=A_ , input_boxes=A_ , return_tensors=A_ , ) return encoding_image_processor def _a ( self , A_ , A_ , A_=None , A_=None , A_=None , A_="pt" , ) -> List[Any]: if input_points is not None: if len(A_ ) != len(A_ ): __UpperCamelCase =[ self._normalize_coordinates(self.target_size , A_ , original_sizes[0] ) for point in input_points ] else: __UpperCamelCase =[ self._normalize_coordinates(self.target_size , A_ , A_ ) for point, original_size in zip(A_ , A_ ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: __UpperCamelCase , __UpperCamelCase =self._pad_points_and_labels(A_ , A_ ) __UpperCamelCase =np.array(A_ ) if input_labels is not None: __UpperCamelCase =np.array(A_ ) if input_boxes is not None: if len(A_ ) != len(A_ ): __UpperCamelCase =[ self._normalize_coordinates(self.target_size , A_ , original_sizes[0] , is_bounding_box=A_ ) for box in input_boxes ] else: __UpperCamelCase =[ self._normalize_coordinates(self.target_size , A_ , A_ , is_bounding_box=A_ ) for box, original_size in zip(A_ , A_ ) ] __UpperCamelCase =np.array(A_ ) if input_boxes is not None: if return_tensors == "pt": __UpperCamelCase =torch.from_numpy(A_ ) # boxes batch size of 1 by default __UpperCamelCase =input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": __UpperCamelCase =tf.convert_to_tensor(A_ ) # boxes batch size of 1 by default __UpperCamelCase =tf.expand_dims(A_ , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'input_boxes': input_boxes} ) if input_points is not None: if return_tensors == "pt": __UpperCamelCase =torch.from_numpy(A_ ) # point batch size of 1 by default __UpperCamelCase =input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": __UpperCamelCase =tf.convert_to_tensor(A_ ) # point batch size of 1 by default __UpperCamelCase =tf.expand_dims(A_ , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'input_points': input_points} ) if input_labels is not None: if return_tensors == "pt": __UpperCamelCase =torch.from_numpy(A_ ) # point batch size of 1 by default __UpperCamelCase =input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": __UpperCamelCase =tf.convert_to_tensor(A_ ) # point batch size of 1 by default __UpperCamelCase =tf.expand_dims(A_ , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'input_labels': input_labels} ) return encoding_image_processor def _a ( self , A_ , A_ ) -> Union[str, Any]: __UpperCamelCase =max([point.shape[0] for point in input_points] ) __UpperCamelCase =[] for i, point in enumerate(A_ ): if point.shape[0] != expected_nb_points: __UpperCamelCase =np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) __UpperCamelCase =np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(A_ ) __UpperCamelCase =processed_input_points return input_points, input_labels def _a ( self , A_ , A_ , A_ , A_=False ) -> np.ndarray: __UpperCamelCase , __UpperCamelCase =original_size __UpperCamelCase , __UpperCamelCase =self.image_processor._get_preprocess_shape(A_ , longest_edge=A_ ) __UpperCamelCase =deepcopy(A_ ).astype(A_ ) if is_bounding_box: __UpperCamelCase =coords.reshape(-1 , 2 , 2 ) __UpperCamelCase =coords[..., 0] * (new_w / old_w) __UpperCamelCase =coords[..., 1] * (new_h / old_h) if is_bounding_box: __UpperCamelCase =coords.reshape(-1 , 4 ) return coords def _a ( self , A_=None , A_=None , A_=None , ) -> List[str]: if input_points is not None: if hasattr(A_ , 'numpy' ): # Checks for TF or Torch tensor __UpperCamelCase =input_points.numpy().tolist() if not isinstance(A_ , A_ ) or not isinstance(input_points[0] , A_ ): raise ValueError('Input points must be a list of list of floating points.' ) __UpperCamelCase =[np.array(A_ ) for input_point in input_points] else: __UpperCamelCase =None if input_labels is not None: if hasattr(A_ , 'numpy' ): __UpperCamelCase =input_labels.numpy().tolist() if not isinstance(A_ , A_ ) or not isinstance(input_labels[0] , A_ ): raise ValueError('Input labels must be a list of list integers.' ) __UpperCamelCase =[np.array(A_ ) for label in input_labels] else: __UpperCamelCase =None if input_boxes is not None: if hasattr(A_ , 'numpy' ): __UpperCamelCase =input_boxes.numpy().tolist() if ( not isinstance(A_ , A_ ) or not isinstance(input_boxes[0] , A_ ) or not isinstance(input_boxes[0][0] , A_ ) ): raise ValueError('Input boxes must be a list of list of list of floating points.' ) __UpperCamelCase =[np.array(A_ ).astype(np.floataa ) for box in input_boxes] else: __UpperCamelCase =None return input_points, input_labels, input_boxes @property def _a ( self ) -> int: __UpperCamelCase =self.image_processor.model_input_names return list(dict.fromkeys(A_ ) ) def _a ( self , *A_ , **A_ ) -> Optional[int]: return self.image_processor.post_process_masks(*A_ , **A_ )
117
from ..utils import DummyObject, requires_backends class UpperCAmelCase__ ( metaclass=A_ ): """simple docstring""" UpperCAmelCase__ : Any = ["speech"] def __init__( self , *A_ , **A_ ) -> Any: requires_backends(self , ['speech'] ) class UpperCAmelCase__ ( metaclass=A_ ): """simple docstring""" UpperCAmelCase__ : Union[str, Any] = ["speech"] def __init__( self , *A_ , **A_ ) -> Union[str, Any]: requires_backends(self , ['speech'] )
117
1