code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' import math def lowerCamelCase (_SCREAMING_SNAKE_CASE : int ): __a : int = math.loga(math.sqrt(4 * positive_integer + 1 ) / 2 + 1 / 2 ) return exponent == int(_SCREAMING_SNAKE_CASE ) def lowerCamelCase (_SCREAMING_SNAKE_CASE : float = 1 / 12_345 ): __a : List[str] = 0 __a : Any = 0 __a : int = 3 while True: __a : Tuple = (integer**2 - 1) / 4 # if candidate is an integer, then there is a partition for k if partition_candidate == int(_SCREAMING_SNAKE_CASE ): __a : str = int(_SCREAMING_SNAKE_CASE ) total_partitions += 1 if check_partition_perfect(_SCREAMING_SNAKE_CASE ): perfect_partitions += 1 if perfect_partitions > 0: if perfect_partitions / total_partitions < max_proportion: return int(_SCREAMING_SNAKE_CASE ) integer += 1 if __name__ == "__main__": print(f'''{solution() = }''')
27
'''simple docstring''' from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __UpperCamelCase ( lowerCAmelCase_ ): def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__a , 'embed_dim' ) ) self.parent.assertTrue(hasattr(__a , 'num_heads' ) ) class __UpperCamelCase : def __init__( self , __a , __a=13 , __a=64 , __a=3 , __a=[16, 48, 96] , __a=[1, 3, 6] , __a=[1, 2, 10] , __a=[7, 3, 3] , __a=[4, 2, 2] , __a=[2, 1, 1] , __a=[2, 2, 2] , __a=[False, False, True] , __a=[0.0, 0.0, 0.0] , __a=0.02 , __a=1E-1_2 , __a=True , __a=True , __a=2 , ): '''simple docstring''' __a : str = parent __a : List[Any] = batch_size __a : Optional[int] = image_size __a : List[str] = patch_sizes __a : str = patch_stride __a : Any = patch_padding __a : Dict = is_training __a : Union[str, Any] = use_labels __a : Dict = num_labels __a : List[Any] = num_channels __a : Any = embed_dim __a : int = num_heads __a : Optional[int] = stride_kv __a : Dict = depth __a : List[str] = cls_token __a : List[Any] = attention_drop_rate __a : Tuple = initializer_range __a : int = layer_norm_eps def __UpperCAmelCase ( self ): '''simple docstring''' __a : str = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a : Dict = None if self.use_labels: # create a random int32 tensor of given shape __a : str = ids_tensor([self.batch_size] , self.num_labels ) __a : str = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): '''simple docstring''' return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def __UpperCAmelCase ( self , __a , __a , __a ): '''simple docstring''' __a : Optional[int] = TFCvtModel(config=__a ) __a : Dict = model(__a , training=__a ) __a : Any = (self.image_size, self.image_size) __a , __a : Dict = image_size[0], image_size[1] for i in range(len(self.depth ) ): __a : Tuple = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __a : str = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def __UpperCAmelCase ( self , __a , __a , __a ): '''simple docstring''' __a : List[Any] = self.num_labels __a : Optional[int] = TFCvtForImageClassification(__a ) __a : Dict = model(__a , labels=__a , training=__a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Optional[Any] = self.prepare_config_and_inputs() __a , __a , __a : Tuple = config_and_inputs __a : str = {'pixel_values': pixel_values} return config, inputs_dict @require_tf class __UpperCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): A_ = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () A_ = ( {"feature-extraction": TFCvtModel, "image-classification": TFCvtForImageClassification} if is_tf_available() else {} ) A_ = False A_ = False A_ = False A_ = False A_ = False def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = TFCvtModelTester(self ) __a : List[Any] = TFCvtConfigTester(self , config_class=__a , has_text_modality=__a , hidden_size=37 ) def __UpperCAmelCase ( self ): '''simple docstring''' self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason='Cvt does not output attentions' ) def __UpperCAmelCase ( self ): '''simple docstring''' pass @unittest.skip(reason='Cvt does not use inputs_embeds' ) def __UpperCAmelCase ( self ): '''simple docstring''' pass @unittest.skip(reason='Cvt does not support input and output embeddings' ) def __UpperCAmelCase ( self ): '''simple docstring''' pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU' ) ) == 0 , reason='TF does not support backprop for grouped convolutions on CPU.' , ) def __UpperCAmelCase ( self ): '''simple docstring''' super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU' ) ) == 0 , reason='TF does not support backprop for grouped convolutions on CPU.' , ) @slow def __UpperCAmelCase ( self ): '''simple docstring''' super().test_keras_fit() @unittest.skip(reason='Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8' ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Any = tf.keras.mixed_precision.Policy('mixed_float16' ) tf.keras.mixed_precision.set_global_policy(__a ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy('float32' ) def __UpperCAmelCase ( self ): '''simple docstring''' __a , __a : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : Dict = model_class(__a ) __a : Optional[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a : Optional[Any] = [*signature.parameters.keys()] __a : Optional[Any] = ['pixel_values'] self.assertListEqual(arg_names[:1] , __a ) def __UpperCAmelCase ( self ): '''simple docstring''' def check_hidden_states_output(__a , __a , __a ): __a : List[str] = model_class(__a ) __a : Union[str, Any] = model(**self._prepare_for_class(__a , __a ) ) __a : Any = outputs.hidden_states __a : Union[str, Any] = len(self.model_tester.depth ) self.assertEqual(len(__a ) , __a ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) __a , __a : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a : List[str] = True check_hidden_states_output(__a , __a , __a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __a : Optional[Any] = True check_hidden_states_output(__a , __a , __a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__a ) def __UpperCAmelCase ( self ): '''simple docstring''' __a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__a ) @slow def __UpperCAmelCase ( self ): '''simple docstring''' for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a : Optional[Any] = TFCvtModel.from_pretrained(__a ) self.assertIsNotNone(__a ) def lowerCamelCase (): __a : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __UpperCamelCase ( unittest.TestCase ): @cached_property def __UpperCAmelCase ( self ): '''simple docstring''' return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __UpperCAmelCase ( self ): '''simple docstring''' __a : int = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __a : Tuple = self.default_image_processor __a : Any = prepare_img() __a : int = image_processor(images=__a , return_tensors='tf' ) # forward pass __a : Any = model(**__a ) # verify the logits __a : Any = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , __a ) __a : Optional[Any] = tf.constant([0.9285, 0.9015, -0.3150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , __a , atol=1E-4 ) )
27
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(SCREAMING_SNAKE_CASE__ ) , 'Tatoeba directory does not exist.' ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCamelCase ( self ) -> str: '''simple docstring''' __UpperCamelCase : Optional[int] = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCamelCase ) @slow def __lowerCamelCase ( self ) -> Any: '''simple docstring''' self.resolver.convert_models(["heb-eng"] ) @slow def __lowerCamelCase ( self ) -> Any: '''simple docstring''' __UpperCamelCase : Any = self.resolver.write_model_card("opus-mt-he-en" , dry_run=__UpperCamelCase ) assert mmeta["long_pair"] == "heb-eng"
359
def UpperCAmelCase_ (_lowerCAmelCase : list ): if len(_lowerCAmelCase ) <= 1: return lst __UpperCamelCase : Dict = 1 while i < len(_lowerCAmelCase ): if lst[i - 1] <= lst[i]: i += 1 else: __UpperCamelCase , __UpperCamelCase : Union[str, Any] = lst[i], lst[i - 1] i -= 1 if i == 0: __UpperCamelCase : Any = 1 return lst if __name__ == "__main__": lowercase : Dict = input("Enter numbers separated by a comma:\n").strip() lowercase : Union[str, Any] = [int(item) for item in user_input.split(",")] print(gnome_sort(unsorted))
171
0
import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def __lowercase ( lowerCamelCase : Dict="" ): UpperCamelCase_ : List[str] = tempfile.mkdtemp() return os.path.join(_a , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class _lowercase ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : int = torch.rand(1_2 , dtype=torch.floataa ) - 0.5 UpperCamelCase_ : Optional[int] = AgentAudio(lowercase_ ) UpperCamelCase_ : List[str] = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowercase_ , agent_type.to_raw() , atol=1e-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(lowercase_ ) ) # Ensure that the file contains the same value as the original tensor UpperCamelCase_ : int = sf.read(lowercase_ ) self.assertTrue(torch.allclose(lowercase_ , torch.tensor(lowercase_ ) , atol=1e-4 ) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Any: """simple docstring""" UpperCamelCase_ : Any = torch.rand(1_2 , dtype=torch.floataa ) - 0.5 UpperCamelCase_ : List[str] = get_new_path(suffix='.wav' ) sf.write(lowercase_ , lowercase_ , 1_6_0_0_0 ) UpperCamelCase_ : Tuple = AgentAudio(lowercase_ ) self.assertTrue(torch.allclose(lowercase_ , agent_type.to_raw() , atol=1e-4 ) ) self.assertEqual(agent_type.to_string() , lowercase_ ) @require_vision @require_torch class _lowercase ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> List[str]: """simple docstring""" UpperCamelCase_ : List[Any] = torch.randint(0 , 2_5_6 , (6_4, 6_4, 3) ) UpperCamelCase_ : str = AgentImage(lowercase_ ) UpperCamelCase_ : Union[str, Any] = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(lowercase_ , agent_type._tensor , atol=1e-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase_ ) ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> List[Any]: """simple docstring""" UpperCamelCase_ : Any = Path(get_tests_dir('fixtures/tests_samples/COCO' ) ) / '''000000039769.png''' UpperCamelCase_ : Optional[int] = Image.open(lowercase_ ) UpperCamelCase_ : Tuple = AgentImage(lowercase_ ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase_ ) ) def SCREAMING_SNAKE_CASE__ ( self : str ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : int = Path(get_tests_dir('fixtures/tests_samples/COCO' ) ) / '''000000039769.png''' UpperCamelCase_ : Dict = Image.open(lowercase_ ) UpperCamelCase_ : List[str] = AgentImage(lowercase_ ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(lowercase_ ) ) class _lowercase ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self : Any ) -> List[str]: """simple docstring""" UpperCamelCase_ : Tuple = '''Hey!''' UpperCamelCase_ : Optional[Any] = AgentText(lowercase_ ) self.assertEqual(lowercase_ , agent_type.to_string() ) self.assertEqual(lowercase_ , agent_type.to_raw() ) self.assertEqual(lowercase_ , lowercase_ )
175
"""simple docstring""" import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() lowercase__ : Dict = [ '''word_embeddings_layernorm.weight''', '''word_embeddings_layernorm.bias''', '''input_layernorm.weight''', '''input_layernorm.bias''', '''post_attention_layernorm.weight''', '''post_attention_layernorm.bias''', '''self_attention.dense.bias''', '''mlp.dense_4h_to_h.bias''', '''ln_f.weight''', '''ln_f.bias''', ] lowercase__ : str = [ '''mlp.dense_4h_to_h.weight''', '''self_attention.dense.weight''', ] def __lowercase ( _a , _a ): snake_case_ : Optional[int] = { '''word_embeddings.weight''': '''word_embeddings.weight''', '''word_embeddings.norm.weight''': '''word_embeddings_layernorm.weight''', '''word_embeddings.norm.bias''': '''word_embeddings_layernorm.bias''', '''weight''': '''ln_f.weight''', '''bias''': '''ln_f.bias''', } if key in layer_rename_map: return layer_rename_map[key] # Handle transformer blocks snake_case_ : List[Any] = int(re.match(r'''.*layer_(\d*).*''' , _a )[1] ) layer_number -= 3 return f"h.{layer_number}." + key def __lowercase ( _a ): if dtype == torch.bool: return 1 / 8 snake_case_ : Dict = re.search(r'''[^\d](\d+)$''' , str(_a ) ) if bit_search is None: raise ValueError(f"`dtype` is not a valid dtype: {dtype}." ) snake_case_ : Optional[int] = int(bit_search.groups()[0] ) return bit_size // 8 def __lowercase ( _a , _a , _a , _a , _a ): # Construct model if bloom_config_file == "": snake_case_ : int = BloomConfig() else: snake_case_ : List[str] = BloomConfig.from_json_file(_a ) if shard_model: snake_case_ : List[str] = os.listdir(_a ) snake_case_ : int = sorted(filter(lambda _a : s.startswith('''layer''' ) and "model_00" in s , _a ) ) snake_case_ : List[str] = {'''weight_map''': {}, '''metadata''': {}} snake_case_ : Any = 0 snake_case_ : Union[str, Any] = None snake_case_ : List[str] = BloomConfig() for j, file in enumerate(_a ): print('''Processing file: {}'''.format(_a ) ) snake_case_ : Dict = None for i in range(_a ): # load all TP files snake_case_ : Union[str, Any] = file.replace('''model_00''' , f"model_0{i}" ) snake_case_ : List[str] = torch.load(os.path.join(_a , _a ) , map_location='''cpu''' ) # Rename keys in the transformers names snake_case_ : str = list(temp.keys() ) for key in keys: snake_case_ : Any = temp.pop(_a ) if tensors is None: snake_case_ : Any = temp else: for key in tensors.keys(): if any(key.endswith(_a ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel snake_case_ : Tuple = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks snake_case_ : List[str] = torch.cat([tensors[key], temp[key]] , dim=_a ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(_a ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): snake_case_ : Any = tensors[key] / pretraining_tp torch.save( _a , os.path.join( _a , '''pytorch_model_{}-of-{}.bin'''.format(str(j + 1 ).zfill(5 ) , str(len(_a ) ).zfill(5 ) ) , ) , ) for key in tensors.keys(): snake_case_ : List[str] = tensors[key] total_size += value.numel() * get_dtype_size(value.dtype ) if key not in index_dict["weight_map"]: snake_case_ : List[str] = '''pytorch_model_{}-of-{}.bin'''.format( str(j + 1 ).zfill(5 ) , str(len(_a ) ).zfill(5 ) ) snake_case_ : int = BloomConfig() snake_case_ : Any = pytorch_dump_folder_path + '''/''' + CONFIG_NAME snake_case_ : Dict = total_size with open(_a , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) with open(os.path.join(_a , WEIGHTS_NAME + '''.index.json''' ) , '''w''' , encoding='''utf-8''' ) as f: snake_case_ : Tuple = json.dumps(_a , indent=2 , sort_keys=_a ) + '''\n''' f.write(_a ) else: snake_case_ : Union[str, Any] = BloomModel(_a ) snake_case_ : List[str] = os.listdir(_a ) snake_case_ : Dict = sorted(filter(lambda _a : s.startswith('''layer''' ) and "model_00" in s , _a ) ) snake_case_ : List[Any] = None for i, file in enumerate(_a ): snake_case_ : Optional[Any] = None for i in range(_a ): # load all TP files snake_case_ : List[str] = file.replace('''model_00''' , f"model_0{i}" ) snake_case_ : Optional[Any] = torch.load(os.path.join(_a , _a ) , map_location='''cpu''' ) # Rename keys in the transformers names snake_case_ : str = list(temp.keys() ) for key in keys: snake_case_ : str = temp.pop(_a ) if tensors is None: snake_case_ : int = temp else: for key in tensors.keys(): # We average (sum and then divide) some weights accross TP ranks (see https://github.com/bigscience-workshop/Megatron-DeepSpeed/blob/olruwase/sync_layer_norms/megatron/training.py#L425) if any(key.endswith(_a ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): tensors[key] += temp[key] else: # Some weights are RowParallelLinear in Megatron-Deepspeed, others are ColumnParallel snake_case_ : Tuple = 1 if any(text in key for text in WEIGHTS_WITH_ROW_PARALLELISM_CONTAIN ) else 0 # We concatenate these weights accross TP ranks snake_case_ : Optional[Any] = torch.cat([tensors[key], temp[key]] , dim=_a ) # Divide by the number of TP the weights we want to average for key in tensors.keys(): if any(key.endswith(_a ) for end in WEIGHTS_TO_AVERAGE_ENDSWITH ): snake_case_ : Union[str, Any] = tensors[key] / pretraining_tp snake_case_ : Any = model.load_state_dict(_a , strict=_a ) assert not other_keys.unexpected_keys, f"The keys {other_keys.unexpected_keys} are unexpected" if missing_keys is None: snake_case_ : Optional[int] = set(other_keys.missing_keys ) else: snake_case_ : Tuple = missing_keys.intersection(set(other_keys.missing_keys ) ) assert not missing_keys, f"The keys {missing_keys} are missing" # Save pytorch-model os.makedirs(_a , exist_ok=_a ) snake_case_ : List[str] = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME snake_case_ : Optional[Any] = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(f"Save PyTorch model to {pytorch_weights_dump_path} with dtype {config.torch_dtype}" ) if config.torch_dtype is not None: snake_case_ : Optional[Any] = model.to(config.torch_dtype ) torch.save(model.state_dict() , _a ) print(f"Save configuration file to {pytorch_config_dump_path}" ) with open(_a , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowercase__ : str = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bloom_checkpoint_path''', default=None, type=str, required=True, help='''Path to the Megatron-LM checkpoint path.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--bloom_config_file''', default='''''', type=str, help=( '''An optional config json file corresponding to the pre-trained model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--shard_model''', action='''store_true''', help='''An optional setting to shard the output model \nThis enables sharding the converted checkpoint''', ) parser.add_argument( '''--pretraining_tp''', default=4, type=int, help='''Pretraining TP rank that has been used when training the model in Megatron-LM \n''', ) lowercase__ : List[Any] = parser.parse_args() convert_bloom_checkpoint_to_pytorch( args.bloom_checkpoint_path, args.bloom_config_file, args.pytorch_dump_folder_path, args.shard_model, args.pretraining_tp, )
264
0
import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SegformerConfig, SegformerForImageClassification, SegformerForSemanticSegmentation, SegformerImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() _lowerCAmelCase : List[str] = logging.get_logger(__name__) def UpperCamelCase_( _snake_case : Tuple , _snake_case : int=False ): """simple docstring""" __a =OrderedDict() for key, value in state_dict.items(): if encoder_only and not key.startswith('head' ): __a ='segformer.encoder.' + key if key.startswith('backbone' ): __a =key.replace('backbone' , 'segformer.encoder' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 __a =key[key.find('patch_embed' ) + len('patch_embed' )] __a =key.replace(F'patch_embed{idx}' , F'patch_embeddings.{int(_snake_case )-1}' ) if "norm" in key: __a =key.replace('norm' , 'layer_norm' ) if "segformer.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 __a =key[key.find('segformer.encoder.layer_norm' ) + len('segformer.encoder.layer_norm' )] __a =key.replace(F'layer_norm{idx}' , F'layer_norm.{int(_snake_case )-1}' ) if "layer_norm1" in key: __a =key.replace('layer_norm1' , 'layer_norm_1' ) if "layer_norm2" in key: __a =key.replace('layer_norm2' , 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 __a =key[key.find('block' ) + len('block' )] __a =key.replace(F'block{idx}' , F'block.{int(_snake_case )-1}' ) if "attn.q" in key: __a =key.replace('attn.q' , 'attention.self.query' ) if "attn.proj" in key: __a =key.replace('attn.proj' , 'attention.output.dense' ) if "attn" in key: __a =key.replace('attn' , 'attention.self' ) if "fc1" in key: __a =key.replace('fc1' , 'dense1' ) if "fc2" in key: __a =key.replace('fc2' , 'dense2' ) if "linear_pred" in key: __a =key.replace('linear_pred' , 'classifier' ) if "linear_fuse" in key: __a =key.replace('linear_fuse.conv' , 'linear_fuse' ) __a =key.replace('linear_fuse.bn' , 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 __a =key[key.find('linear_c' ) + len('linear_c' )] __a =key.replace(F'linear_c{idx}' , F'linear_c.{int(_snake_case )-1}' ) if key.startswith('head' ): __a =key.replace('head' , 'classifier' ) __a =value return new_state_dict def UpperCamelCase_( _snake_case : List[Any] , _snake_case : Union[str, Any] ): """simple docstring""" for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) __a =state_dict.pop(F'segformer.encoder.block.{i}.{j}.attention.self.kv.weight' ) __a =state_dict.pop(F'segformer.encoder.block.{i}.{j}.attention.self.kv.bias' ) # next, add keys and values (in that order) to the state dict __a =kv_weight[ : config.hidden_sizes[i], : ] __a =kv_bias[: config.hidden_sizes[i]] __a =kv_weight[ config.hidden_sizes[i] :, : ] __a =kv_bias[ config.hidden_sizes[i] : ] def UpperCamelCase_( ): """simple docstring""" __a ='http://images.cocodataset.org/val2017/000000039769.jpg' __a =Image.open(requests.get(_snake_case , stream=_snake_case ).raw ) return image @torch.no_grad() def UpperCamelCase_( _snake_case : Optional[int] , _snake_case : Union[str, Any] , _snake_case : Optional[int] ): """simple docstring""" __a =SegformerConfig() __a =False # set attributes based on model_name __a ='huggingface/label-files' if "segformer" in model_name: __a =model_name[len('segformer.' ) : len('segformer.' ) + 2] if "ade" in model_name: __a =150 __a ='ade20k-id2label.json' __a =(1, 150, 128, 128) elif "city" in model_name: __a =19 __a ='cityscapes-id2label.json' __a =(1, 19, 128, 128) else: raise ValueError(F'Model {model_name} not supported' ) elif "mit" in model_name: __a =True __a =model_name[4:6] __a =1000 __a ='imagenet-1k-id2label.json' __a =(1, 1000) else: raise ValueError(F'Model {model_name} not supported' ) # set config attributes __a =json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type='dataset' ) , 'r' ) ) __a ={int(_snake_case ): v for k, v in idalabel.items()} __a =idalabel __a ={v: k for k, v in idalabel.items()} if size == "b0": pass elif size == "b1": __a =[64, 128, 320, 512] __a =256 elif size == "b2": __a =[64, 128, 320, 512] __a =768 __a =[3, 4, 6, 3] elif size == "b3": __a =[64, 128, 320, 512] __a =768 __a =[3, 4, 18, 3] elif size == "b4": __a =[64, 128, 320, 512] __a =768 __a =[3, 8, 27, 3] elif size == "b5": __a =[64, 128, 320, 512] __a =768 __a =[3, 6, 40, 3] else: raise ValueError(F'Size {size} not supported' ) # load image processor (only resize + normalize) __a =SegformerImageProcessor( image_scale=(512, 512) , keep_ratio=_snake_case , align=_snake_case , do_random_crop=_snake_case ) # prepare image __a =prepare_img() __a =image_processor(images=_snake_case , return_tensors='pt' ).pixel_values logger.info(F'Converting model {model_name}...' ) # load original state dict if encoder_only: __a =torch.load(_snake_case , map_location=torch.device('cpu' ) ) else: __a =torch.load(_snake_case , map_location=torch.device('cpu' ) )['state_dict'] # rename keys __a =rename_keys(_snake_case , encoder_only=_snake_case ) if not encoder_only: del state_dict["decode_head.conv_seg.weight"] del state_dict["decode_head.conv_seg.bias"] # key and value matrices need special treatment read_in_k_v(_snake_case , _snake_case ) # create HuggingFace model and load state dict if encoder_only: __a =False __a =SegformerForImageClassification(_snake_case ) else: __a =SegformerForSemanticSegmentation(_snake_case ) model.load_state_dict(_snake_case ) model.eval() # forward pass __a =model(_snake_case ) __a =outputs.logits # set expected_slice based on model name # ADE20k checkpoints if model_name == "segformer.b0.512x512.ade.160k": __a =torch.tensor( [ [[-4.6_310, -5.5_232, -6.2_356], [-5.1_921, -6.1_444, -6.5_996], [-5.4_424, -6.2_790, -6.7_574]], [[-12.1_391, -13.3_122, -13.9_554], [-12.8_732, -13.9_352, -14.3_563], [-12.9_438, -13.8_226, -14.2_513]], [[-12.5_134, -13.4_686, -14.4_915], [-12.8_669, -14.4_343, -14.7_758], [-13.2_523, -14.5_819, -15.0_694]], ] ) elif model_name == "segformer.b1.512x512.ade.160k": __a =torch.tensor( [ [[-7.5_820, -8.7_231, -8.3_215], [-8.0_600, -10.3_529, -10.0_304], [-7.5_208, -9.4_103, -9.6_239]], [[-12.6_918, -13.8_994, -13.7_137], [-13.3_196, -15.7_523, -15.4_789], [-12.9_343, -14.8_757, -14.9_689]], [[-11.1_911, -11.9_421, -11.3_243], [-11.3_342, -13.6_839, -13.3_581], [-10.3_909, -12.1_832, -12.4_858]], ] ) elif model_name == "segformer.b2.512x512.ade.160k": __a =torch.tensor( [ [[-11.8_173, -14.3_850, -16.3_128], [-14.5_648, -16.5_804, -18.6_568], [-14.7_223, -15.7_387, -18.4_218]], [[-15.7_290, -17.9_171, -19.4_423], [-18.3_105, -19.9_448, -21.4_661], [-17.9_296, -18.6_497, -20.7_910]], [[-15.0_783, -17.0_336, -18.2_789], [-16.8_771, -18.6_870, -20.1_612], [-16.2_454, -17.1_426, -19.5_055]], ] ) elif model_name == "segformer.b3.512x512.ade.160k": __a =torch.tensor( [ [[-9.0_878, -10.2_081, -10.1_891], [-9.3_144, -10.7_941, -10.9_843], [-9.2_294, -10.3_855, -10.5_704]], [[-12.2_316, -13.9_068, -13.6_102], [-12.9_161, -14.3_702, -14.3_235], [-12.5_233, -13.7_174, -13.7_932]], [[-14.6_275, -15.2_490, -14.9_727], [-14.3_400, -15.9_687, -16.2_827], [-14.1_484, -15.4_033, -15.8_937]], ] ) elif model_name == "segformer.b4.512x512.ade.160k": __a =torch.tensor( [ [[-12.3_144, -13.2_447, -14.0_802], [-13.3_614, -14.5_816, -15.6_117], [-13.3_340, -14.4_433, -16.2_219]], [[-19.2_781, -20.4_128, -20.7_506], [-20.6_153, -21.6_566, -22.0_998], [-19.9_800, -21.0_430, -22.1_494]], [[-18.8_739, -19.7_804, -21.1_834], [-20.1_233, -21.6_765, -23.2_944], [-20.0_315, -21.2_641, -23.6_944]], ] ) elif model_name == "segformer.b5.640x640.ade.160k": __a =torch.tensor( [ [[-9.5_524, -12.0_835, -11.7_348], [-10.5_229, -13.6_446, -14.5_662], [-9.5_842, -12.8_851, -13.9_414]], [[-15.3_432, -17.5_323, -17.0_818], [-16.3_330, -18.9_255, -19.2_101], [-15.1_340, -17.7_848, -18.3_971]], [[-12.6_072, -14.9_486, -14.6_631], [-13.7_629, -17.0_907, -17.7_745], [-12.7_899, -16.1_695, -17.1_671]], ] ) # Cityscapes checkpoints elif model_name == "segformer.b0.1024x1024.city.160k": __a =torch.tensor( [ [[-11.9_295, -13.4_057, -14.8_106], [-13.3_431, -14.8_179, -15.3_781], [-14.2_836, -15.5_942, -16.1_588]], [[-11.4_906, -12.8_067, -13.6_564], [-13.1_189, -14.0_500, -14.1_543], [-13.8_748, -14.5_136, -14.8_789]], [[0.5_374, 0.1_067, -0.4_742], [0.1_141, -0.2_255, -0.7_099], [-0.3_000, -0.5_924, -1.3_105]], ] ) elif model_name == "segformer.b0.512x1024.city.160k": __a =torch.tensor( [ [[-7.8_217, -9.8_767, -10.1_717], [-9.4_438, -10.9_058, -11.4_047], [-9.7_939, -12.3_495, -12.1_079]], [[-7.1_514, -9.5_336, -10.0_860], [-9.7_776, -11.6_822, -11.8_439], [-10.1_411, -12.7_655, -12.8_972]], [[0.3_021, 0.0_805, -0.2_310], [-0.0_328, -0.1_605, -0.2_714], [-0.1_408, -0.5_477, -0.6_976]], ] ) elif model_name == "segformer.b0.640x1280.city.160k": __a =torch.tensor( [ [ [-1.1_3_7_2e0_1, -1.2_7_8_7e0_1, -1.3_4_7_7e0_1], [-1.2_5_3_6e0_1, -1.4_1_9_4e0_1, -1.4_4_0_9e0_1], [-1.3_2_1_7e0_1, -1.4_8_8_8e0_1, -1.5_3_2_7e0_1], ], [ [-1.4_7_9_1e0_1, -1.7_1_2_2e0_1, -1.8_2_7_7e0_1], [-1.7_1_6_3e0_1, -1.9_1_9_2e0_1, -1.9_5_3_3e0_1], [-1.7_8_9_7e0_1, -1.9_9_9_1e0_1, -2.0_3_1_5e0_1], ], [ [7.6_7_2_3e-0_1, 4.1_9_2_1e-0_1, -7.7_8_7_8e-0_2], [4.7_7_7_2e-0_1, 9.5_5_5_7e-0_3, -2.8_0_8_2e-0_1], [3.6_0_3_2e-0_1, -2.4_8_2_6e-0_1, -5.1_1_6_8e-0_1], ], ] ) elif model_name == "segformer.b0.768x768.city.160k": __a =torch.tensor( [ [[-9.4_959, -11.3_087, -11.7_479], [-11.0_025, -12.6_540, -12.3_319], [-11.4_064, -13.0_487, -12.9_905]], [[-9.8_905, -11.3_084, -12.0_854], [-11.1_726, -12.7_698, -12.9_583], [-11.5_985, -13.3_278, -14.1_774]], [[0.2_213, 0.0_192, -0.2_466], [-0.1_731, -0.4_213, -0.4_874], [-0.3_126, -0.6_541, -1.1_389]], ] ) elif model_name == "segformer.b1.1024x1024.city.160k": __a =torch.tensor( [ [[-13.5_748, -13.9_111, -12.6_500], [-14.3_500, -15.3_683, -14.2_328], [-14.7_532, -16.0_424, -15.6_087]], [[-17.1_651, -15.8_725, -12.9_653], [-17.2_580, -17.3_718, -14.8_223], [-16.6_058, -16.8_783, -16.7_452]], [[-3.6_456, -3.0_209, -1.4_203], [-3.0_797, -3.1_959, -2.0_000], [-1.8_757, -1.9_217, -1.6_997]], ] ) elif model_name == "segformer.b2.1024x1024.city.160k": __a =torch.tensor( [ [[-16.0_976, -16.4_856, -17.3_962], [-16.6_234, -19.0_342, -19.7_685], [-16.0_900, -18.0_661, -19.1_180]], [[-18.4_750, -18.8_488, -19.5_074], [-19.4_030, -22.1_570, -22.5_977], [-19.1_191, -20.8_486, -22.3_783]], [[-4.5_178, -5.5_037, -6.5_109], [-5.0_884, -7.2_174, -8.0_334], [-4.4_156, -5.8_117, -7.2_970]], ] ) elif model_name == "segformer.b3.1024x1024.city.160k": __a =torch.tensor( [ [[-14.2_081, -14.4_732, -14.1_977], [-14.5_867, -16.4_423, -16.6_356], [-13.4_441, -14.9_685, -16.8_696]], [[-14.4_576, -14.7_073, -15.0_451], [-15.0_816, -17.6_237, -17.9_873], [-14.4_213, -16.0_199, -18.5_992]], [[-4.7_349, -4.9_588, -5.0_966], [-4.3_210, -6.9_325, -7.2_591], [-3.4_312, -4.7_484, -7.1_917]], ] ) elif model_name == "segformer.b4.1024x1024.city.160k": __a =torch.tensor( [ [[-11.7_737, -11.9_526, -11.3_273], [-13.6_692, -14.4_574, -13.8_878], [-13.8_937, -14.6_924, -15.9_345]], [[-14.6_706, -14.5_330, -14.1_306], [-16.1_502, -16.8_180, -16.4_269], [-16.8_338, -17.8_939, -20.1_746]], [[1.0_491, 0.8_289, 1.0_310], [1.1_044, 0.5_219, 0.8_055], [1.0_899, 0.6_926, 0.5_590]], ] ) elif model_name == "segformer.b5.1024x1024.city.160k": __a =torch.tensor( [ [[-12.5_641, -13.4_777, -13.0_684], [-13.9_587, -15.8_983, -16.6_557], [-13.3_109, -15.7_350, -16.3_141]], [[-14.7_074, -15.4_352, -14.5_944], [-16.6_353, -18.1_663, -18.6_120], [-15.1_702, -18.0_329, -18.1_547]], [[-1.7_990, -2.0_951, -1.7_784], [-2.6_397, -3.8_245, -3.9_686], [-1.5_264, -2.8_126, -2.9_316]], ] ) else: __a =logits.argmax(-1 ).item() print('Predicted class:' , model.config.idalabel[predicted_class_idx] ) # verify logits if not encoder_only: assert logits.shape == expected_shape assert torch.allclose(logits[0, :3, :3, :3] , _snake_case , atol=1e-2 ) # finally, save model and image processor logger.info(F'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(_snake_case ).mkdir(exist_ok=_snake_case ) model.save_pretrained(_snake_case ) image_processor.save_pretrained(_snake_case ) if __name__ == "__main__": _lowerCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument( "--model_name", default="segformer.b0.512x512.ade.160k", type=str, help="Name of the model you'd like to convert.", ) parser.add_argument( "--checkpoint_path", default=None, type=str, help="Path to the original PyTorch checkpoint (.pth file)." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the folder to output PyTorch model." ) _lowerCAmelCase : Union[str, Any] = parser.parse_args() convert_segformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
362
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _lowerCAmelCase : int = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase : Dict = ["BartphoTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _lowerCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
308
0
'''simple docstring''' import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_logger _SCREAMING_SNAKE_CASE = get_logger(__name__) class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase = None ) -> Union[str, Any]: _lowerCAmelCase = ( os.path.join(_UpperCAmelCase , config.EXTRACTED_DATASETS_DIR ) if cache_dir else config.EXTRACTED_DATASETS_PATH ) _lowerCAmelCase = Extractor def _snake_case ( self , _lowerCAmelCase ) -> Tuple: from .file_utils import hash_url_to_filename # Path where we extract compressed archives # We extract in the cache dir, and get the extracted path name by hashing the original path" _lowerCAmelCase = os.path.abspath(_UpperCAmelCase ) return os.path.join(self.extract_dir , hash_url_to_filename(_UpperCAmelCase ) ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase ) -> Any: return force_extract or ( not os.path.isfile(_UpperCAmelCase ) and not (os.path.isdir(_UpperCAmelCase ) and os.listdir(_UpperCAmelCase )) ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase = False ) -> int: _lowerCAmelCase = self.extractor.infer_extractor_format(_UpperCAmelCase ) if not extractor_format: return input_path _lowerCAmelCase = self._get_output_path(_UpperCAmelCase ) if self._do_extract(_UpperCAmelCase , _UpperCAmelCase ): self.extractor.extract(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return output_path class lowerCAmelCase_ ( _UpperCamelCase ): @classmethod @abstractmethod def _snake_case ( cls , _lowerCAmelCase , **_lowerCAmelCase ) -> int: ... @staticmethod @abstractmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: ... class lowerCAmelCase_ ( _UpperCamelCase ,_UpperCamelCase ): __lowerCamelCase : int = [] @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> List[str]: with open(_UpperCAmelCase , "rb" ) as f: return f.read(_UpperCAmelCase ) @classmethod def _snake_case ( cls , _lowerCAmelCase , _lowerCAmelCase = b"" ) -> List[Any]: if not magic_number: _lowerCAmelCase = max(len(_UpperCAmelCase ) for cls_magic_number in cls.magic_numbers ) try: _lowerCAmelCase = cls.read_magic_number(_UpperCAmelCase , _UpperCAmelCase ) except OSError: return False return any(magic_number.startswith(_UpperCAmelCase ) for cls_magic_number in cls.magic_numbers ) class lowerCAmelCase_ ( _UpperCamelCase ): @classmethod def _snake_case ( cls , _lowerCAmelCase , **_lowerCAmelCase ) -> Optional[int]: return tarfile.is_tarfile(_UpperCAmelCase ) @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> Any: def resolved(_lowerCAmelCase ) -> str: return os.path.realpath(os.path.abspath(_UpperCAmelCase ) ) def badpath(_lowerCAmelCase , _lowerCAmelCase ) -> bool: # joinpath will ignore base if path is absolute return not resolved(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ).startswith(_UpperCAmelCase ) def badlink(_lowerCAmelCase , _lowerCAmelCase ) -> bool: # Links are interpreted relative to the directory containing the link _lowerCAmelCase = resolved(os.path.join(_UpperCAmelCase , os.path.dirname(info.name ) ) ) return badpath(info.linkname , base=_UpperCAmelCase ) _lowerCAmelCase = resolved(_UpperCAmelCase ) for finfo in members: if badpath(finfo.name , _UpperCAmelCase ): logger.error(f'''Extraction of {finfo.name} is blocked (illegal path)''' ) elif finfo.issym() and badlink(_UpperCAmelCase , _UpperCAmelCase ): logger.error(f'''Extraction of {finfo.name} is blocked: Symlink to {finfo.linkname}''' ) elif finfo.islnk() and badlink(_UpperCAmelCase , _UpperCAmelCase ): logger.error(f'''Extraction of {finfo.name} is blocked: Hard link to {finfo.linkname}''' ) else: yield finfo @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> Dict: os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) _lowerCAmelCase = tarfile.open(_UpperCAmelCase ) tar_file.extractall(_UpperCAmelCase , members=TarExtractor.safemembers(_UpperCAmelCase , _UpperCAmelCase ) ) tar_file.close() class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : Tuple = [B"\x1F\x8B"] @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> str: with gzip.open(_UpperCAmelCase , "rb" ) as gzip_file: with open(_UpperCAmelCase , "wb" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : Union[str, Any] = [ B"PK\x03\x04", B"PK\x05\x06", # empty archive B"PK\x07\x08", # spanned archive ] @classmethod def _snake_case ( cls , _lowerCAmelCase , _lowerCAmelCase = b"" ) -> str: if super().is_extractable(_UpperCAmelCase , magic_number=_UpperCAmelCase ): return True try: # Alternative version of zipfile.is_zipfile that has less false positives, but misses executable zip archives. # From: https://github.com/python/cpython/pull/5053 from zipfile import ( _CD_SIGNATURE, _ECD_DISK_NUMBER, _ECD_DISK_START, _ECD_ENTRIES_TOTAL, _ECD_OFFSET, _ECD_SIZE, _EndRecData, sizeCentralDir, stringCentralDir, structCentralDir, ) with open(_UpperCAmelCase , "rb" ) as fp: _lowerCAmelCase = _EndRecData(_UpperCAmelCase ) if endrec: if endrec[_ECD_ENTRIES_TOTAL] == 0 and endrec[_ECD_SIZE] == 0 and endrec[_ECD_OFFSET] == 0: return True # Empty zipfiles are still zipfiles elif endrec[_ECD_DISK_NUMBER] == endrec[_ECD_DISK_START]: fp.seek(endrec[_ECD_OFFSET] ) # Central directory is on the same disk if fp.tell() == endrec[_ECD_OFFSET] and endrec[_ECD_SIZE] >= sizeCentralDir: _lowerCAmelCase = fp.read(_UpperCAmelCase ) # CD is where we expect it to be if len(_UpperCAmelCase ) == sizeCentralDir: _lowerCAmelCase = struct.unpack(_UpperCAmelCase , _UpperCAmelCase ) # CD is the right size if centdir[_CD_SIGNATURE] == stringCentralDir: return True # First central directory entry has correct magic number return False except Exception: # catch all errors in case future python versions change the zipfile internals return False @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> int: os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) with zipfile.ZipFile(_UpperCAmelCase , "r" ) as zip_file: zip_file.extractall(_UpperCAmelCase ) zip_file.close() class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : Any = [B"\xFD\x37\x7A\x58\x5A\x00"] @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> int: with lzma.open(_UpperCAmelCase ) as compressed_file: with open(_UpperCAmelCase , "wb" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : List[Any] = [B"Rar!\x1a\x07\x00", B"Rar!\x1a\x07\x01\x00"] # RAR_ID # RAR5_ID @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: if not config.RARFILE_AVAILABLE: raise ImportError("Please pip install rarfile" ) import rarfile os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) _lowerCAmelCase = rarfile.RarFile(_UpperCAmelCase ) rf.extractall(_UpperCAmelCase ) rf.close() class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : Tuple = [B"\x28\xb5\x2F\xFD"] @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[int]: if not config.ZSTANDARD_AVAILABLE: raise ImportError("Please pip install zstandard" ) import zstandard as zstd _lowerCAmelCase = zstd.ZstdDecompressor() with open(_UpperCAmelCase , "rb" ) as ifh, open(_UpperCAmelCase , "wb" ) as ofh: dctx.copy_stream(_UpperCAmelCase , _UpperCAmelCase ) class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : Any = [B"\x42\x5A\x68"] @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> int: with bza.open(_UpperCAmelCase , "rb" ) as compressed_file: with open(_UpperCAmelCase , "wb" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : List[str] = [B"\x37\x7A\xBC\xAF\x27\x1C"] @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: if not config.PY7ZR_AVAILABLE: raise ImportError("Please pip install py7zr" ) import pyazr os.makedirs(_UpperCAmelCase , exist_ok=_UpperCAmelCase ) with pyazr.SevenZipFile(_UpperCAmelCase , "r" ) as archive: archive.extractall(_UpperCAmelCase ) class lowerCAmelCase_ ( _UpperCamelCase ): __lowerCamelCase : Optional[int] = [B"\x04\x22\x4D\x18"] @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: if not config.LZ4_AVAILABLE: raise ImportError("Please pip install lz4" ) import lza.frame with lza.frame.open(_UpperCAmelCase , "rb" ) as compressed_file: with open(_UpperCAmelCase , "wb" ) as extracted_file: shutil.copyfileobj(_UpperCAmelCase , _UpperCAmelCase ) class lowerCAmelCase_ : __lowerCamelCase : int = { "tar": TarExtractor, "gzip": GzipExtractor, "zip": ZipExtractor, "xz": XzExtractor, "rar": RarExtractor, "zstd": ZstdExtractor, "bz2": BzipaExtractor, "7z": SevenZipExtractor, # <Added version="2.4.0"/> "lz4": LzaExtractor, # <Added version="2.4.0"/> } @classmethod def _snake_case ( cls ) -> Optional[Any]: return max( len(_UpperCAmelCase ) for extractor in cls.extractors.values() if issubclass(_UpperCAmelCase , _UpperCAmelCase ) for extractor_magic_number in extractor.magic_numbers ) @staticmethod def _snake_case ( _lowerCAmelCase , _lowerCAmelCase ) -> int: try: return MagicNumberBaseExtractor.read_magic_number(_UpperCAmelCase , magic_number_length=_UpperCAmelCase ) except OSError: return b"" @classmethod def _snake_case ( cls , _lowerCAmelCase , _lowerCAmelCase = False ) -> List[str]: warnings.warn( "Method \'is_extractable\' was deprecated in version 2.4.0 and will be removed in 3.0.0. " "Use \'infer_extractor_format\' instead." , category=_UpperCAmelCase , ) _lowerCAmelCase = cls.infer_extractor_format(_UpperCAmelCase ) if extractor_format: return True if not return_extractor else (True, cls.extractors[extractor_format]) return False if not return_extractor else (False, None) @classmethod def _snake_case ( cls , _lowerCAmelCase ) -> Tuple: # <Added version="2.4.0"/> _lowerCAmelCase = cls._get_magic_number_max_length() _lowerCAmelCase = cls._read_magic_number(_UpperCAmelCase , _UpperCAmelCase ) for extractor_format, extractor in cls.extractors.items(): if extractor.is_extractable(_UpperCAmelCase , magic_number=_UpperCAmelCase ): return extractor_format @classmethod def _snake_case ( cls , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = "deprecated" , ) -> Any: os.makedirs(os.path.dirname(_UpperCAmelCase ) , exist_ok=_UpperCAmelCase ) # Prevent parallel extractions _lowerCAmelCase = str(Path(_UpperCAmelCase ).with_suffix(".lock" ) ) with FileLock(_UpperCAmelCase ): shutil.rmtree(_UpperCAmelCase , ignore_errors=_UpperCAmelCase ) if extractor_format or extractor != "deprecated": if extractor != "deprecated" or not isinstance(_UpperCAmelCase , _UpperCAmelCase ): # passed as positional arg warnings.warn( "Parameter \'extractor\' was deprecated in version 2.4.0 and will be removed in 3.0.0. " "Use \'extractor_format\' instead." , category=_UpperCAmelCase , ) _lowerCAmelCase = extractor if extractor != '''deprecated''' else extractor_format else: _lowerCAmelCase = cls.extractors[extractor_format] return extractor.extract(_UpperCAmelCase , _UpperCAmelCase ) else: warnings.warn( "Parameter \'extractor_format\' was made required in version 2.4.0 and not passing it will raise an " "exception in 3.0.0." , category=_UpperCAmelCase , ) for extractor in cls.extractors.values(): if extractor.is_extractable(_UpperCAmelCase ): return extractor.extract(_UpperCAmelCase , _UpperCAmelCase )
158
"""simple docstring""" import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml A = NewType('''DataClass''', Any) A = NewType('''DataClassType''', Any) def __A ( a_ :List[str]) -> Tuple: if isinstance(a_ , a_): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( F"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""") def __A ( a_ :list) -> Callable[[str], Any]: __a : Any = {str(a_): choice for choice in choices} return lambda a_: str_to_choice.get(a_ , a_) def __A ( *, a_ :Union[str, List[str]] = None , a_ :str = None , a_ :Any = dataclasses.MISSING , a_ :Callable[[], Any] = dataclasses.MISSING , a_ :dict = None , **a_ :str , ) -> dataclasses.Field: if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls __a : List[Any] = {} if aliases is not None: __a : Optional[Any] = aliases if help is not None: __a : int = help return dataclasses.field(metadata=a_ , default=a_ , default_factory=a_ , **a_) class __lowercase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = 42 def __init__( self , _UpperCAmelCase , **_UpperCAmelCase ): # To make the default appear when using --help if "formatter_class" not in kwargs: __a : str = ArgumentDefaultsHelpFormatter super().__init__(**_UpperCAmelCase ) if dataclasses.is_dataclass(_UpperCAmelCase ): __a : int = [dataclass_types] __a : Optional[Any] = list(_UpperCAmelCase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(_UpperCAmelCase ) @staticmethod def _lowerCamelCase ( _UpperCAmelCase , _UpperCAmelCase ): __a : List[Any] = f"""--{field.name}""" __a : Optional[int] = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , _UpperCAmelCase ): raise RuntimeError( '''Unresolved type detected, which should have been done with the help of ''' '''`typing.get_type_hints` method by default''' ) __a : Dict = kwargs.pop('''aliases''' , [] ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): __a : List[str] = [aliases] __a : Tuple = getattr(field.type , '''__origin__''' , field.type ) if origin_type is Union or (hasattr(_UpperCAmelCase , '''UnionType''' ) and isinstance(_UpperCAmelCase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(_UpperCAmelCase ) not in field.type.__args__ ): raise ValueError( '''Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because''' ''' the argument parser only supports one type per argument.''' f""" Problem encountered in field '{field.name}'.""" ) if type(_UpperCAmelCase ) not in field.type.__args__: # filter `str` in Union __a : List[str] = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] __a : List[str] = getattr(field.type , '''__origin__''' , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) __a : List[str] = ( field.type.__args__[0] if isinstance(_UpperCAmelCase , field.type.__args__[1] ) else field.type.__args__[1] ) __a : Optional[Any] = getattr(field.type , '''__origin__''' , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) __a : Optional[int] = {} if origin_type is Literal or (isinstance(field.type , _UpperCAmelCase ) and issubclass(field.type , _UpperCAmelCase )): if origin_type is Literal: __a : int = field.type.__args__ else: __a : List[str] = [x.value for x in field.type] __a : Any = make_choice_type_function(kwargs['''choices'''] ) if field.default is not dataclasses.MISSING: __a : Tuple = field.default else: __a : Optional[int] = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument __a : Any = copy(_UpperCAmelCase ) # Hack because type=bool in argparse does not behave as we want. __a : List[str] = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. __a : Union[str, Any] = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way __a : List[Any] = default # This tells argparse we accept 0 or 1 value after --field_name __a : Union[str, Any] = '''?''' # This is the value that will get picked if we do --field_name (without value) __a : List[Any] = True elif isclass(_UpperCAmelCase ) and issubclass(_UpperCAmelCase , _UpperCAmelCase ): __a : Dict = field.type.__args__[0] __a : Optional[int] = '''+''' if field.default_factory is not dataclasses.MISSING: __a : Union[str, Any] = field.default_factory() elif field.default is dataclasses.MISSING: __a : List[Any] = True else: __a : int = field.type if field.default is not dataclasses.MISSING: __a : Optional[Any] = field.default elif field.default_factory is not dataclasses.MISSING: __a : Optional[int] = field.default_factory() else: __a : Union[str, Any] = True parser.add_argument(_UpperCAmelCase , *_UpperCAmelCase , **_UpperCAmelCase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): __a : Any = False parser.add_argument(f"""--no_{field.name}""" , action='''store_false''' , dest=field.name , **_UpperCAmelCase ) def _lowerCamelCase ( self , _UpperCAmelCase ): if hasattr(_UpperCAmelCase , '''_argument_group_name''' ): __a : Any = self.add_argument_group(dtype._argument_group_name ) else: __a : Optional[Any] = self try: __a : Dict[str, type] = get_type_hints(_UpperCAmelCase ) except NameError: raise RuntimeError( f"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ '''removing line of `from __future__ import annotations` which opts in Postponed ''' '''Evaluation of Annotations (PEP 563)''' ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(_UpperCAmelCase ): __a : Union[str, Any] = '''.'''.join(map(_UpperCAmelCase , sys.version_info[:3] ) ) raise RuntimeError( f"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ '''line of `from __future__ import annotations` which opts in union types as ''' '''`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To ''' '''support Python versions that lower than 3.10, you need to use ''' '''`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of ''' '''`X | None`.''' ) from ex raise for field in dataclasses.fields(_UpperCAmelCase ): if not field.init: continue __a : str = type_hints[field.name] self._parse_dataclass_field(_UpperCAmelCase , _UpperCAmelCase ) def _lowerCamelCase ( self , _UpperCAmelCase=None , _UpperCAmelCase=False , _UpperCAmelCase=True , _UpperCAmelCase=None , _UpperCAmelCase=None , ): if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): __a : int = [] if args_filename: args_files.append(Path(_UpperCAmelCase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix('''.args''' ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values __a : Optional[Any] = ArgumentParser() args_file_parser.add_argument(_UpperCAmelCase , type=_UpperCAmelCase , action='''append''' ) # Use only remaining args for further parsing (remove the args_file_flag) __a , __a : List[Any] = args_file_parser.parse_known_args(args=_UpperCAmelCase ) __a : Union[str, Any] = vars(_UpperCAmelCase ).get(args_file_flag.lstrip('''-''' ) , _UpperCAmelCase ) if cmd_args_file_paths: args_files.extend([Path(_UpperCAmelCase ) for p in cmd_args_file_paths] ) __a : Union[str, Any] = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last __a : Dict = file_args + args if args is not None else file_args + sys.argv[1:] __a , __a : str = self.parse_known_args(args=_UpperCAmelCase ) __a : Optional[int] = [] for dtype in self.dataclass_types: __a : Optional[int] = {f.name for f in dataclasses.fields(_UpperCAmelCase ) if f.init} __a : List[str] = {k: v for k, v in vars(_UpperCAmelCase ).items() if k in keys} for k in keys: delattr(_UpperCAmelCase , _UpperCAmelCase ) __a : int = dtype(**_UpperCAmelCase ) outputs.append(_UpperCAmelCase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(_UpperCAmelCase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(f"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase = False ): __a : Tuple = set(args.keys() ) __a : List[str] = [] for dtype in self.dataclass_types: __a : Dict = {f.name for f in dataclasses.fields(_UpperCAmelCase ) if f.init} __a : Union[str, Any] = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) __a : Tuple = dtype(**_UpperCAmelCase ) outputs.append(_UpperCAmelCase ) if not allow_extra_keys and unused_keys: raise ValueError(f"""Some keys are not used by the HfArgumentParser: {sorted(_UpperCAmelCase )}""" ) return tuple(_UpperCAmelCase ) def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase = False ): with open(Path(_UpperCAmelCase ) , encoding='''utf-8''' ) as open_json_file: __a : int = json.loads(open_json_file.read() ) __a : str = self.parse_dict(_UpperCAmelCase , allow_extra_keys=_UpperCAmelCase ) return tuple(_UpperCAmelCase ) def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase = False ): __a : Tuple = self.parse_dict(yaml.safe_load(Path(_UpperCAmelCase ).read_text() ) , allow_extra_keys=_UpperCAmelCase ) return tuple(_UpperCAmelCase )
160
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: UpperCAmelCase = None UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} UpperCAmelCase = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", }, """tokenizer_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json""", }, } UpperCAmelCase = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } UpperCAmelCase = """▁""" class lowerCAmelCase_ ( lowerCamelCase__ ): '''simple docstring''' __snake_case = VOCAB_FILES_NAMES __snake_case = PRETRAINED_VOCAB_FILES_MAP __snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case = AlbertTokenizer def __init__( self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=True , _UpperCAmelCase=True , _UpperCAmelCase=False , _UpperCAmelCase="[CLS]" , _UpperCAmelCase="[SEP]" , _UpperCAmelCase="<unk>" , _UpperCAmelCase="[SEP]" , _UpperCAmelCase="<pad>" , _UpperCAmelCase="[CLS]" , _UpperCAmelCase="[MASK]" , **_UpperCAmelCase , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. snake_case_ = ( AddedToken(_UpperCAmelCase , lstrip=_UpperCAmelCase , rstrip=_UpperCAmelCase , normalized=_UpperCAmelCase ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ) else mask_token ) super().__init__( _UpperCAmelCase , tokenizer_file=_UpperCAmelCase , do_lower_case=_UpperCAmelCase , remove_space=_UpperCAmelCase , keep_accents=_UpperCAmelCase , bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , unk_token=_UpperCAmelCase , sep_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , cls_token=_UpperCAmelCase , mask_token=_UpperCAmelCase , **_UpperCAmelCase , ) snake_case_ = do_lower_case snake_case_ = remove_space snake_case_ = keep_accents snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase = None ): snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase = None ): snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase__ ( self , _UpperCAmelCase , _UpperCAmelCase = None ): if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_UpperCAmelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return snake_case_ = os.path.join( _UpperCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCAmelCase ): copyfile(self.vocab_file , _UpperCAmelCase ) return (out_vocab_file,)
267
import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase = get_tests_dir("""fixtures/test_sentencepiece_no_bos.model""") @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' __snake_case = PegasusTokenizer __snake_case = PegasusTokenizerFast __snake_case = True __snake_case = True def UpperCamelCase__ ( self ): super().setUp() # We have a SentencePiece fixture for testing snake_case_ = PegasusTokenizer(_UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase__ ( self ): return PegasusTokenizer.from_pretrained('''google/pegasus-large''' ) def UpperCamelCase__ ( self , **_UpperCAmelCase ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase ): return ("This is a test", "This is a test") def UpperCamelCase__ ( self ): snake_case_ = '''</s>''' snake_case_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_UpperCAmelCase ) , _UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_UpperCAmelCase ) , _UpperCAmelCase ) def UpperCamelCase__ ( self ): snake_case_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<pad>''' ) self.assertEqual(vocab_keys[1] , '''</s>''' ) self.assertEqual(vocab_keys[-1] , '''v''' ) self.assertEqual(len(_UpperCAmelCase ) , 11_03 ) def UpperCamelCase__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 11_03 ) def UpperCamelCase__ ( self ): snake_case_ = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = self.tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = ( '''Let\'s see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important''' ''' </s> <pad> <pad> <pad>''' ) snake_case_ = rust_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] snake_case_ = py_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase__ ( self ): snake_case_ = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word snake_case_ = '''<mask_1> To ensure a <mask_2> flow of bank resolutions.''' snake_case_ = [2, 4_13, 6_15, 1_14, 3, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] snake_case_ = tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) def UpperCamelCase__ ( self ): snake_case_ = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 9_61_03 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 1_03 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 1_05 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 10_24 snake_case_ = '''To ensure a smooth flow of bank resolutions.''' snake_case_ = [4_13, 6_15, 1_14, 22_91, 19_71, 1_13, 16_79, 1_07_10, 1_07, 1] snake_case_ = tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def UpperCamelCase__ ( self ): snake_case_ = ['''This is going to be way too long.''' * 1_50, '''short example'''] snake_case_ = ['''not super long but more than 5 tokens''', '''tiny'''] snake_case_ = self._large_tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) snake_case_ = self._large_tokenizer( text_target=_UpperCAmelCase , max_length=5 , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 10_24) assert batch.attention_mask.shape == (2, 10_24) assert targets["input_ids"].shape == (2, 5) assert len(_UpperCAmelCase ) == 2 # input_ids, attention_mask. @slow def UpperCamelCase__ ( self ): # fmt: off snake_case_ = {'''input_ids''': [[3_89_79, 1_43, 1_84_85, 6_06, 1_30, 2_66_69, 8_76_86, 1_21, 5_41_89, 11_29, 1_11, 2_66_69, 8_76_86, 1_21, 91_14, 1_47_87, 1_21, 1_32_49, 1_58, 5_92, 9_56, 1_21, 1_46_21, 3_15_76, 1_43, 6_26_13, 1_08, 96_88, 9_30, 4_34_30, 1_15_62, 6_26_13, 3_04, 1_08, 1_14_43, 8_97, 1_08, 93_14, 1_74_15, 6_33_99, 1_08, 1_14_43, 76_14, 1_83_16, 1_18, 42_84, 71_48, 1_24_30, 1_43, 14_00, 2_57_03, 1_58, 1_11, 42_84, 71_48, 1_17_72, 1_43, 2_12_97, 10_64, 1_58, 1_22, 2_04, 35_06, 17_54, 11_33, 1_47_87, 15_81, 1_15, 3_32_24, 44_82, 1_11, 13_55, 1_10, 2_91_73, 3_17, 5_08_33, 1_08, 2_01_47, 9_46_65, 1_11, 7_71_98, 1_07, 1], [1_10, 6_26_13, 1_17, 6_38, 1_12, 11_33, 1_21, 2_00_98, 13_55, 7_90_50, 1_38_72, 1_35, 15_96, 5_35_41, 13_52, 1_41, 1_30_39, 55_42, 1_24, 3_02, 5_18, 1_11, 2_68, 29_56, 1_15, 1_49, 44_27, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_39, 12_35, 27_99, 1_82_89, 1_77_80, 2_04, 1_09, 94_74, 12_96, 1_07, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_UpperCAmelCase , model_name='''google/bigbird-pegasus-large-arxiv''' , revision='''ba85d0851d708441f91440d509690f1ab6353415''' , ) @require_sentencepiece @require_tokenizers class lowerCAmelCase_ ( lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' __snake_case = PegasusTokenizer __snake_case = PegasusTokenizerFast __snake_case = True __snake_case = True def UpperCamelCase__ ( self ): super().setUp() # We have a SentencePiece fixture for testing snake_case_ = PegasusTokenizer(_UpperCAmelCase , offset=0 , mask_token_sent=_UpperCAmelCase , mask_token='''[MASK]''' ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def UpperCamelCase__ ( self ): return PegasusTokenizer.from_pretrained('''google/bigbird-pegasus-large-arxiv''' ) def UpperCamelCase__ ( self , **_UpperCAmelCase ): return PegasusTokenizer.from_pretrained(self.tmpdirname , **_UpperCAmelCase ) def UpperCamelCase__ ( self , _UpperCAmelCase ): return ("This is a test", "This is a test") def UpperCamelCase__ ( self ): snake_case_ = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = self.tokenizer_class.from_pretrained(self.tmpdirname ) snake_case_ = ( '''Let\'s see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>''' ''' <pad> <pad> <pad>''' ) snake_case_ = rust_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] snake_case_ = py_tokenizer([raw_input_str] , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase ).input_ids[0] self.assertListEqual(_UpperCAmelCase , _UpperCAmelCase ) @require_torch def UpperCamelCase__ ( self ): snake_case_ = ['''This is going to be way too long.''' * 10_00, '''short example'''] snake_case_ = ['''not super long but more than 5 tokens''', '''tiny'''] snake_case_ = self._large_tokenizer(_UpperCAmelCase , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) snake_case_ = self._large_tokenizer( text_target=_UpperCAmelCase , max_length=5 , padding=_UpperCAmelCase , truncation=_UpperCAmelCase , return_tensors='''pt''' ) assert batch.input_ids.shape == (2, 40_96) assert batch.attention_mask.shape == (2, 40_96) assert targets["input_ids"].shape == (2, 5) assert len(_UpperCAmelCase ) == 2 # input_ids, attention_mask. def UpperCamelCase__ ( self ): snake_case_ = ( '''This is an example string that is used to test the original TF implementation against the HF''' ''' implementation''' ) snake_case_ = self._large_tokenizer(_UpperCAmelCase ).input_ids self.assertListEqual( _UpperCAmelCase , [1_82, 1_17, 1_42, 5_87, 42_11, 1_20, 1_17, 2_63, 1_12, 8_04, 1_09, 8_56, 2_50_16, 31_37, 4_64, 1_09, 2_69_55, 31_37, 1] , )
267
1
"""simple docstring""" import os import shutil import sys import tempfile import unittest from pathlib import Path import pytest import transformers from transformers import ( BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoTokenizer, BertConfig, BertTokenizer, BertTokenizerFast, CTRLTokenizer, GPTaTokenizer, GPTaTokenizerFast, PreTrainedTokenizerFast, RobertaTokenizer, RobertaTokenizerFast, is_tokenizers_available, ) from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.auto.tokenization_auto import ( TOKENIZER_MAPPING, get_tokenizer_config, tokenizer_class_from_name, ) from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import ( DUMMY_DIFF_TOKENIZER_IDENTIFIER, DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tokenizers, slow, ) sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self :Any ) -> Tuple: UpperCAmelCase = 0 @slow def UpperCAmelCase__ ( self :Dict ) -> int: for model_name in (x for x in BERT_PRETRAINED_CONFIG_ARCHIVE_MAP.keys() if "japanese" not in x): UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , (BertTokenizer, BertTokenizerFast) ) self.assertGreater(len(lowercase_ ) , 0 ) for model_name in GPT2_PRETRAINED_CONFIG_ARCHIVE_MAP.keys(): UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) self.assertIsInstance(lowercase_ , (GPTaTokenizer, GPTaTokenizerFast) ) self.assertGreater(len(lowercase_ ) , 0 ) def UpperCAmelCase__ ( self :List[Any] ) -> List[Any]: UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def UpperCAmelCase__ ( self :Dict ) -> str: UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , (RobertaTokenizer, RobertaTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 20 ) def UpperCAmelCase__ ( self :Optional[int] ) -> int: UpperCAmelCase = AutoConfig.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) # Check that tokenizer_type ≠ model_type UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , config=lowercase_ ) self.assertIsInstance(lowercase_ , (BertTokenizer, BertTokenizerFast) ) self.assertEqual(tokenizer.vocab_size , 12 ) def UpperCAmelCase__ ( self :Any ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.txt' , os.path.join(lowercase_ , 'vocab.txt' ) ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , tokenizer_type='bert' , use_fast=lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.json' , os.path.join(lowercase_ , 'vocab.json' ) ) shutil.copy('./tests/fixtures/merges.txt' , os.path.join(lowercase_ , 'merges.txt' ) ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , tokenizer_type='gpt2' , use_fast=lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) @require_tokenizers def UpperCAmelCase__ ( self :Union[str, Any] ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.txt' , os.path.join(lowercase_ , 'vocab.txt' ) ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , tokenizer_type='bert' ) self.assertIsInstance(lowercase_ , lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: shutil.copy('./tests/fixtures/vocab.json' , os.path.join(lowercase_ , 'vocab.json' ) ) shutil.copy('./tests/fixtures/merges.txt' , os.path.join(lowercase_ , 'merges.txt' ) ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , tokenizer_type='gpt2' ) self.assertIsInstance(lowercase_ , lowercase_ ) def UpperCAmelCase__ ( self :Any ) -> List[Any]: with pytest.raises(lowercase_ ): AutoTokenizer.from_pretrained('./' , tokenizer_type='xxx' ) @require_tokenizers def UpperCAmelCase__ ( self :List[str] ) -> List[str]: for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: UpperCAmelCase = tokenizer_class.from_pretrained('wietsedv/bert-base-dutch-cased' ) self.assertIsInstance(lowercase_ , (BertTokenizer, BertTokenizerFast) ) if isinstance(lowercase_ , lowercase_ ): self.assertEqual(tokenizer.basic_tokenizer.do_lower_case , lowercase_ ) else: self.assertEqual(tokenizer.do_lower_case , lowercase_ ) self.assertEqual(tokenizer.model_max_length , 5_12 ) @require_tokenizers def UpperCAmelCase__ ( self :str ) -> Any: for tokenizer_class in [BertTokenizer, BertTokenizerFast, AutoTokenizer]: with self.assertRaisesRegex( lowercase_ , 'julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier' , ): UpperCAmelCase = tokenizer_class.from_pretrained('julien-c/herlolip-not-exists' ) def UpperCAmelCase__ ( self :Any ) -> Optional[Any]: # tests: https://github.com/huggingface/transformers/pull/13251 # 1. models with `-`, e.g. xlm-roberta -> xlm_roberta # 2. models that don't remap 1-1 from model-name to model file, e.g., openai-gpt -> openai UpperCAmelCase = TOKENIZER_MAPPING.values() UpperCAmelCase = [] for slow_tok, fast_tok in tokenizers: if slow_tok is not None: tokenizer_names.append(slow_tok.__name__ ) if fast_tok is not None: tokenizer_names.append(fast_tok.__name__ ) for tokenizer_name in tokenizer_names: # must find the right class tokenizer_class_from_name(lowercase_ ) @require_tokenizers def UpperCAmelCase__ ( self :Dict ) -> Dict: self.assertIsInstance(AutoTokenizer.from_pretrained('bert-base-cased' , use_fast=lowercase_ ) , lowercase_ ) self.assertIsInstance(AutoTokenizer.from_pretrained('bert-base-cased' ) , lowercase_ ) @require_tokenizers def UpperCAmelCase__ ( self :Union[str, Any] ) -> str: UpperCAmelCase = AutoTokenizer.from_pretrained('distilbert-base-uncased' , do_lower_case=lowercase_ ) UpperCAmelCase = 'Hello, world. How are you?' UpperCAmelCase = tokenizer.tokenize(lowercase_ ) self.assertEqual('[UNK]' , tokens[0] ) UpperCAmelCase = AutoTokenizer.from_pretrained('microsoft/mpnet-base' , do_lower_case=lowercase_ ) UpperCAmelCase = tokenizer.tokenize(lowercase_ ) self.assertEqual('[UNK]' , tokens[0] ) @require_tokenizers def UpperCAmelCase__ ( self :int ) -> str: UpperCAmelCase = AutoTokenizer.from_pretrained('robot-test/dummy-tokenizer-fast-with-model-config' ) self.assertEqual(type(lowercase_ ) , lowercase_ ) self.assertEqual(tokenizer.model_max_length , 5_12 ) self.assertEqual(tokenizer.vocab_size , 3_00_00 ) self.assertEqual(tokenizer.unk_token , '[UNK]' ) self.assertEqual(tokenizer.padding_side , 'right' ) self.assertEqual(tokenizer.truncation_side , 'right' ) def UpperCAmelCase__ ( self :int ) -> Tuple: UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , (BertTokenizer, BertTokenizerFast) ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , tokenizer.__class__ ) self.assertEqual(tokenizera.vocab_size , 12 ) def UpperCAmelCase__ ( self :Dict ) -> Optional[int]: UpperCAmelCase = AutoTokenizer.from_pretrained('ctrl' ) # There is no fast CTRL so this always gives us a slow tokenizer. self.assertIsInstance(lowercase_ , lowercase_ ) def UpperCAmelCase__ ( self :str ) -> Optional[int]: # Check we can load the tokenizer config of an online model. UpperCAmelCase = get_tokenizer_config('bert-base-cased' ) UpperCAmelCase = config.pop('_commit_hash' , lowercase_ ) # If we ever update bert-base-cased tokenizer config, this dict here will need to be updated. self.assertEqual(lowercase_ , {'do_lower_case': False} ) # This model does not have a tokenizer_config so we get back an empty dict. UpperCAmelCase = get_tokenizer_config(lowercase_ ) self.assertDictEqual(lowercase_ , {} ) # A tokenizer saved with `save_pretrained` always creates a tokenizer config. UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = get_tokenizer_config(lowercase_ ) # Check the class of the tokenizer was properly saved (note that it always saves the slow class). self.assertEqual(config['tokenizer_class'] , 'BertTokenizer' ) def UpperCAmelCase__ ( self :Dict ) -> Tuple: try: AutoConfig.register('custom' , lowercase_ ) AutoTokenizer.register(lowercase_ , slow_tokenizer_class=lowercase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase_ ): AutoTokenizer.register(lowercase_ , slow_tokenizer_class=lowercase_ ) UpperCAmelCase = CustomTokenizer.from_pretrained(lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] @require_tokenizers def UpperCAmelCase__ ( self :Tuple ) -> List[Any]: try: AutoConfig.register('custom' , lowercase_ ) # Can register in two steps AutoTokenizer.register(lowercase_ , slow_tokenizer_class=lowercase_ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, None) ) AutoTokenizer.register(lowercase_ , fast_tokenizer_class=lowercase_ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) del TOKENIZER_MAPPING._extra_content[CustomConfig] # Can register in one step AutoTokenizer.register( lowercase_ , slow_tokenizer_class=lowercase_ , fast_tokenizer_class=lowercase_ ) self.assertEqual(TOKENIZER_MAPPING[CustomConfig] , (CustomTokenizer, CustomTokenizerFast) ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowercase_ ): AutoTokenizer.register(lowercase_ , fast_tokenizer_class=lowercase_ ) # We pass through a bert tokenizer fast cause there is no converter slow to fast for our new toknizer # and that model does not have a tokenizer.json with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase = BertTokenizerFast.from_pretrained(lowercase_ ) bert_tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = CustomTokenizerFast.from_pretrained(lowercase_ ) with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , use_fast=lowercase_ ) self.assertIsInstance(lowercase_ , lowercase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def UpperCAmelCase__ ( self :Dict ) -> Tuple: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(lowercase_ ): UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowercase_ ): UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=lowercase_ ) self.assertTrue(tokenizer.special_attribute_present ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , trust_remote_code=lowercase_ ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=lowercase_ , use_fast=lowercase_ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) # Test tokenizer can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(lowercase_ ) UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , trust_remote_code=lowercase_ , use_fast=lowercase_ ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertTrue(reloaded_tokenizer.special_attribute_present ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertEqual(reloaded_tokenizer.__class__.__name__ , 'NewTokenizer' ) @require_tokenizers def UpperCAmelCase__ ( self :int ) -> Optional[int]: class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = False class A_ ( SCREAMING_SNAKE_CASE_ ): """simple docstring""" __UpperCamelCase = NewTokenizer __UpperCamelCase = False try: AutoConfig.register('custom' , lowercase_ ) AutoTokenizer.register(lowercase_ , slow_tokenizer_class=lowercase_ ) AutoTokenizer.register(lowercase_ , fast_tokenizer_class=lowercase_ ) # If remote code is not set, the default is to use local UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertFalse(tokenizer.special_attribute_present ) UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/test_dynamic_tokenizer' , use_fast=lowercase_ ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertFalse(tokenizer.special_attribute_present ) # If remote code is disabled, we load the local one. UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=lowercase_ ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertFalse(tokenizer.special_attribute_present ) UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=lowercase_ , use_fast=lowercase_ ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertFalse(tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=lowercase_ ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) self.assertTrue(tokenizer.special_attribute_present ) UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer' , trust_remote_code=lowercase_ , use_fast=lowercase_ ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) self.assertTrue(tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] def UpperCAmelCase__ ( self :Tuple ) -> List[str]: UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer_legacy' , trust_remote_code=lowercase_ ) self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version UpperCAmelCase = AutoTokenizer.from_pretrained( 'hf-internal-testing/test_dynamic_tokenizer_legacy' , trust_remote_code=lowercase_ , use_fast=lowercase_ ) self.assertTrue(tokenizer.special_attribute_present ) self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) def UpperCAmelCase__ ( self :List[Any] ) -> List[Any]: with self.assertRaisesRegex( lowercase_ , 'bert-base is not a local folder and is not a valid model identifier' ): UpperCAmelCase = AutoTokenizer.from_pretrained('bert-base' ) def UpperCAmelCase__ ( self :List[Any] ) -> Optional[int]: with self.assertRaisesRegex( lowercase_ , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): UpperCAmelCase = AutoTokenizer.from_pretrained(lowercase_ , revision='aaaaaa' ) def UpperCAmelCase__ ( self :Tuple ) -> Tuple: # Make sure we have cached the tokenizer. UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) with RequestCounter() as counter: UpperCAmelCase = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
78
"""simple docstring""" from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
78
1
"""simple docstring""" import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class UpperCamelCase : """simple docstring""" def __init__( self ,UpperCAmelCase_ ,UpperCAmelCase_=2 ,UpperCAmelCase_=32 ,UpperCAmelCase_=16 ,UpperCAmelCase_=3 ,UpperCAmelCase_=True ,UpperCAmelCase_=True ,UpperCAmelCase_=32 ,UpperCAmelCase_=4 ,UpperCAmelCase_=[0, 1, 2, 3] ,UpperCAmelCase_=4 ,UpperCAmelCase_=37 ,UpperCAmelCase_="gelu" ,UpperCAmelCase_=0.1 ,UpperCAmelCase_=0.1 ,UpperCAmelCase_=0.02 ,UpperCAmelCase_=3 ,UpperCAmelCase_=[1, 3_84, 24, 24] ,UpperCAmelCase_=True ,UpperCAmelCase_=None ,): _lowercase : Dict = parent _lowercase : List[str] = batch_size _lowercase : Optional[Any] = image_size _lowercase : Tuple = patch_size _lowercase : Optional[int] = num_channels _lowercase : Optional[int] = is_training _lowercase : Dict = use_labels _lowercase : int = hidden_size _lowercase : Tuple = num_hidden_layers _lowercase : Dict = backbone_out_indices _lowercase : Tuple = num_attention_heads _lowercase : List[Any] = intermediate_size _lowercase : Optional[Any] = hidden_act _lowercase : Union[str, Any] = hidden_dropout_prob _lowercase : Optional[Any] = attention_probs_dropout_prob _lowercase : List[Any] = initializer_range _lowercase : str = num_labels _lowercase : Optional[int] = backbone_featmap_shape _lowercase : Tuple = scope _lowercase : int = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) _lowercase : Optional[int] = (image_size // patch_size) ** 2 _lowercase : str = num_patches + 1 def lowerCamelCase__ ( self ): _lowercase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowercase : Union[str, Any] = None if self.use_labels: _lowercase : Union[str, Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] ,self.num_labels ) _lowercase : Any = self.get_config() return config, pixel_values, labels def lowerCamelCase__ ( self ): _lowercase : Any = { """global_padding""": """same""", """layer_type""": """bottleneck""", """depths""": [3, 4, 9], """out_features""": ["""stage1""", """stage2""", """stage3"""], """embedding_dynamic_padding""": True, """hidden_sizes""": [96, 1_92, 3_84, 7_68], """num_groups""": 2, } return DPTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,backbone_out_indices=self.backbone_out_indices ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=UpperCAmelCase_ ,initializer_range=self.initializer_range ,is_hybrid=self.is_hybrid ,backbone_config=UpperCAmelCase_ ,backbone_featmap_shape=self.backbone_featmap_shape ,) def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ): _lowercase : Optional[int] = DPTModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _lowercase : List[str] = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ): _lowercase : str = self.num_labels _lowercase : str = DPTForDepthEstimation(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _lowercase : int = model(UpperCAmelCase_ ) self.parent.assertEqual(result.predicted_depth.shape ,(self.batch_size, self.image_size, self.image_size) ) def lowerCamelCase__ ( self ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ): _lowercase : Tuple = self.num_labels _lowercase : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() _lowercase : int = model(UpperCAmelCase_ ,labels=UpperCAmelCase_ ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_labels, self.image_size, self.image_size) ) def lowerCamelCase__ ( self ): _lowercase : Any = self.prepare_config_and_inputs() _lowercase : Optional[int] = config_and_inputs _lowercase : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class UpperCamelCase ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () SCREAMING_SNAKE_CASE_ : Tuple = ( { "depth-estimation": DPTForDepthEstimation, "feature-extraction": DPTModel, "image-segmentation": DPTForSemanticSegmentation, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ : Optional[int] = False SCREAMING_SNAKE_CASE_ : List[Any] = False SCREAMING_SNAKE_CASE_ : int = False def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = DPTModelTester(self ) _lowercase : Tuple = ConfigTester(self ,config_class=UpperCAmelCase_ ,has_text_modality=UpperCAmelCase_ ,hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""DPT does not use inputs_embeds""" ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): _lowercase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : List[str] = model_class(UpperCAmelCase_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) _lowercase : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(UpperCAmelCase_ ,nn.Linear ) ) def lowerCamelCase__ ( self ): _lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowercase : int = model_class(UpperCAmelCase_ ) _lowercase : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowercase : List[str] = [*signature.parameters.keys()] _lowercase : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] ,UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*UpperCAmelCase_ ) def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCAmelCase_ ) def lowerCamelCase__ ( self ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue _lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : Optional[int] = True if model_class in get_values(UpperCAmelCase_ ): continue _lowercase : Tuple = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.train() _lowercase : Dict = self._prepare_for_class(UpperCAmelCase_ ,UpperCAmelCase_ ,return_labels=UpperCAmelCase_ ) _lowercase : Optional[int] = model(**UpperCAmelCase_ ).loss loss.backward() def lowerCamelCase__ ( self ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue _lowercase : int = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : str = False _lowercase : Any = True if model_class in get_values(UpperCAmelCase_ ) or not model_class.supports_gradient_checkpointing: continue _lowercase : List[str] = model_class(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.gradient_checkpointing_enable() model.train() _lowercase : Optional[int] = self._prepare_for_class(UpperCAmelCase_ ,UpperCAmelCase_ ,return_labels=UpperCAmelCase_ ) _lowercase : Optional[Any] = model(**UpperCAmelCase_ ).loss loss.backward() def lowerCamelCase__ ( self ): _lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : str = _config_zero_init(UpperCAmelCase_ ) for model_class in self.all_model_classes: _lowercase : Optional[int] = model_class(config=UpperCAmelCase_ ) # Skip the check for the backbone _lowercase : Dict = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": _lowercase : Optional[Any] = [f"""{name}.{key}""" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() ,[0.0, 1.0] ,msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" ,) @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def lowerCamelCase__ ( self ): pass @slow def lowerCamelCase__ ( self ): for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: _lowercase : Dict = DPTModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) def lowerCamelCase__ ( self ): # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type _lowercase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _lowercase : Dict = """add""" with self.assertRaises(UpperCAmelCase_ ): _lowercase : Tuple = DPTForDepthEstimation(UpperCAmelCase_ ) def __SCREAMING_SNAKE_CASE ( ): _lowercase : Union[str, Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision @slow class UpperCamelCase ( unittest.TestCase ): """simple docstring""" def lowerCamelCase__ ( self ): _lowercase : Optional[Any] = DPTImageProcessor.from_pretrained("""Intel/dpt-hybrid-midas""" ) _lowercase : Optional[int] = DPTForDepthEstimation.from_pretrained("""Intel/dpt-hybrid-midas""" ).to(UpperCAmelCase_ ) _lowercase : Tuple = prepare_img() _lowercase : Dict = image_processor(images=UpperCAmelCase_ ,return_tensors="""pt""" ).to(UpperCAmelCase_ ) # forward pass with torch.no_grad(): _lowercase : List[str] = model(**UpperCAmelCase_ ) _lowercase : Dict = outputs.predicted_depth # verify the predicted depth _lowercase : Tuple = torch.Size((1, 3_84, 3_84) ) self.assertEqual(predicted_depth.shape ,UpperCAmelCase_ ) _lowercase : Any = torch.tensor( [[[5.6437, 5.6146, 5.6511], [5.4371, 5.5649, 5.5958], [5.5215, 5.5184, 5.5293]]] ).to(UpperCAmelCase_ ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 1_00 ,UpperCAmelCase_ ,atol=1E-4 ) )
351
"""simple docstring""" import re from filelock import FileLock try: import nltk UpperCAmelCase: List[str] = True except (ImportError, ModuleNotFoundError): UpperCAmelCase: int = False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punkt""", quiet=True) def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): re.sub("""<n>""" , """""" , __UpperCAmelCase ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(__UpperCAmelCase ) )
336
0
'''simple docstring''' import math def lowercase ( __magic_name__ ): '''simple docstring''' UpperCAmelCase : Optional[int] = [True] * n UpperCAmelCase : Tuple = False UpperCAmelCase : str = False UpperCAmelCase : Any = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): UpperCAmelCase : Union[str, Any] = i * 2 while index < n: UpperCAmelCase : Dict = False UpperCAmelCase : int = index + i UpperCAmelCase : Union[str, Any] = [2] for i in range(3 , __magic_name__ , 2 ): if is_prime[i]: primes.append(__magic_name__ ) return primes def lowercase ( __magic_name__ = 9999_6666_3333 ): '''simple docstring''' UpperCAmelCase : Union[str, Any] = math.floor(math.sqrt(__magic_name__ ) ) + 100 UpperCAmelCase : List[Any] = prime_sieve(__magic_name__ ) UpperCAmelCase : Tuple = 0 UpperCAmelCase : str = 0 UpperCAmelCase : Optional[Any] = primes[prime_index] while (last_prime**2) <= limit: UpperCAmelCase : Optional[int] = primes[prime_index + 1] UpperCAmelCase : Union[str, Any] = last_prime**2 UpperCAmelCase : List[Any] = next_prime**2 # Get numbers divisible by lps(current) UpperCAmelCase : int = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) UpperCAmelCase : List[str] = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps UpperCAmelCase : Union[str, Any] = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair UpperCAmelCase : List[Any] = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
311
'''simple docstring''' def lowercase ( __magic_name__ ): '''simple docstring''' if number > 0: raise ValueError("input must be a negative integer" ) UpperCAmelCase : List[Any] = len(bin(__magic_name__ )[3:] ) UpperCAmelCase : Optional[Any] = bin(abs(__magic_name__ ) - (1 << binary_number_length) )[3:] UpperCAmelCase : Tuple = ( ( "1" + "0" * (binary_number_length - len(__magic_name__ )) + twos_complement_number ) if number < 0 else "0" ) return "0b" + twos_complement_number if __name__ == "__main__": import doctest doctest.testmod()
311
1
from __future__ import annotations def lowerCamelCase_ ( _a : dict , _a : str ): '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ : int = set(_a ), [start] while stack: UpperCAmelCase_ : Optional[int] = stack.pop() explored.add(_a ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(_a ) return explored UpperCamelCase_ = { '''A''': ['''B''', '''C''', '''D'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F'''], '''D''': ['''B''', '''D'''], '''E''': ['''B''', '''F'''], '''F''': ['''C''', '''E''', '''G'''], '''G''': ['''F'''], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, '''A'''))
59
import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() UpperCamelCase_ = logging.get_logger('''transformers.models.speecht5''') def lowerCamelCase_ ( _a : str , _a : int , _a : Union[str, Any] ): '''simple docstring''' hf_model.apply_weight_norm() UpperCAmelCase_ : Optional[int] = checkpoint["""input_conv.weight_g"""] UpperCAmelCase_ : str = checkpoint["""input_conv.weight_v"""] UpperCAmelCase_ : str = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): UpperCAmelCase_ : Dict = checkpoint[F'''upsamples.{i}.1.weight_g'''] UpperCAmelCase_ : Any = checkpoint[F'''upsamples.{i}.1.weight_v'''] UpperCAmelCase_ : Union[str, Any] = checkpoint[F'''upsamples.{i}.1.bias'''] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): UpperCAmelCase_ : Tuple = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_g'''] UpperCAmelCase_ : Dict = checkpoint[F'''blocks.{i}.convs1.{j}.1.weight_v'''] UpperCAmelCase_ : Optional[Any] = checkpoint[F'''blocks.{i}.convs1.{j}.1.bias'''] UpperCAmelCase_ : Tuple = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_g'''] UpperCAmelCase_ : Optional[Any] = checkpoint[F'''blocks.{i}.convs2.{j}.1.weight_v'''] UpperCAmelCase_ : Tuple = checkpoint[F'''blocks.{i}.convs2.{j}.1.bias'''] UpperCAmelCase_ : Optional[int] = checkpoint["""output_conv.1.weight_g"""] UpperCAmelCase_ : Optional[Any] = checkpoint["""output_conv.1.weight_v"""] UpperCAmelCase_ : Union[str, Any] = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def lowerCamelCase_ ( _a : Tuple , _a : int , _a : Any , _a : Tuple=None , _a : Dict=None , ): '''simple docstring''' if config_path is not None: UpperCAmelCase_ : Union[str, Any] = SpeechTaHifiGanConfig.from_pretrained(_a ) else: UpperCAmelCase_ : str = SpeechTaHifiGanConfig() UpperCAmelCase_ : List[str] = SpeechTaHifiGan(_a ) UpperCAmelCase_ : int = torch.load(_a ) load_weights(orig_checkpoint["""model"""]["""generator"""] , _a , _a ) UpperCAmelCase_ : List[Any] = np.load(_a ) UpperCAmelCase_ : Optional[Any] = stats[0].reshape(-1 ) UpperCAmelCase_ : int = stats[1].reshape(-1 ) UpperCAmelCase_ : Any = torch.from_numpy(_a ).float() UpperCAmelCase_ : int = torch.from_numpy(_a ).float() model.save_pretrained(_a ) if repo_id: print("""Pushing to the hub...""" ) model.push_to_hub(_a ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--stats_path''', required=True, default=None, type=str, help='''Path to stats.npy file''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) UpperCamelCase_ = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
59
1
"""simple docstring""" import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , __UpperCAmelCase , __UpperCAmelCase=1_3 , __UpperCAmelCase=3_2 , __UpperCAmelCase=3 , __UpperCAmelCase=4 , __UpperCAmelCase=[1_0, 2_0, 3_0, 4_0] , __UpperCAmelCase=[2, 2, 3, 2] , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=3_7 , __UpperCAmelCase="gelu" , __UpperCAmelCase=1_0 , __UpperCAmelCase=0.02 , __UpperCAmelCase=["stage2", "stage3", "stage4"] , __UpperCAmelCase=[2, 3, 4] , __UpperCAmelCase=None , ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = parent lowerCAmelCase__ :Dict = batch_size lowerCAmelCase__ :List[str] = image_size lowerCAmelCase__ :str = num_channels lowerCAmelCase__ :str = num_stages lowerCAmelCase__ :Dict = hidden_sizes lowerCAmelCase__ :Tuple = depths lowerCAmelCase__ :Any = is_training lowerCAmelCase__ :Optional[int] = use_labels lowerCAmelCase__ :Optional[Any] = intermediate_size lowerCAmelCase__ :List[str] = hidden_act lowerCAmelCase__ :Optional[int] = num_labels lowerCAmelCase__ :str = initializer_range lowerCAmelCase__ :List[Any] = out_features lowerCAmelCase__ :Union[str, Any] = out_indices lowerCAmelCase__ :Any = scope def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ :List[str] = None if self.use_labels: lowerCAmelCase__ :Tuple = ids_tensor([self.batch_size] , self.num_labels ) lowerCAmelCase__ :List[Any] = self.get_config() return config, pixel_values, labels def snake_case ( self ): '''simple docstring''' return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Any = ConvNextVaModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowerCAmelCase__ :Dict = model(lowerCAmelCase__ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = ConvNextVaForImageClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowerCAmelCase__ :int = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' lowerCAmelCase__ :List[str] = ConvNextVaBackbone(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowerCAmelCase__ :Dict = model(lowerCAmelCase__ ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None lowerCAmelCase__ :Dict = None lowerCAmelCase__ :Any = ConvNextVaBackbone(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() lowerCAmelCase__ :int = model(lowerCAmelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Any = self.prepare_config_and_inputs() lowerCAmelCase__ :Union[str, Any] = config_and_inputs lowerCAmelCase__ :List[str] = {"""pixel_values""": pixel_values} return config, inputs_dict def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :str = self.prepare_config_and_inputs() lowerCAmelCase__ :Optional[Any] = config_and_inputs lowerCAmelCase__ :Tuple = {"""pixel_values""": pixel_values, """labels""": labels} return config, inputs_dict @require_torch class _lowerCAmelCase ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" __magic_name__ :Any = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) __magic_name__ :Optional[int] = ( {'''feature-extraction''': ConvNextVaModel, '''image-classification''': ConvNextVaForImageClassification} if is_torch_available() else {} ) __magic_name__ :Union[str, Any] = False __magic_name__ :int = False __magic_name__ :List[str] = False __magic_name__ :List[str] = False __magic_name__ :List[str] = False def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Union[str, Any] = ConvNextVaModelTester(self ) lowerCAmelCase__ :Optional[int] = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ , hidden_size=3_7 ) def snake_case ( self ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case ( self ): '''simple docstring''' return @unittest.skip(reason='ConvNextV2 does not use inputs_embeds' ) def snake_case ( self ): '''simple docstring''' pass @unittest.skip(reason='ConvNextV2 does not support input and output embeddings' ) def snake_case ( self ): '''simple docstring''' pass @unittest.skip(reason='ConvNextV2 does not use feedforward chunking' ) def snake_case ( self ): '''simple docstring''' pass def snake_case ( self ): '''simple docstring''' if not self.model_tester.is_training: return for model_class in self.all_model_classes: lowerCAmelCase__ :Dict = self.model_tester.prepare_config_and_inputs_with_labels() lowerCAmelCase__ :List[Any] = True if model_class.__name__ in [ *get_values(lowerCAmelCase__ ), *get_values(lowerCAmelCase__ ), ]: continue lowerCAmelCase__ :Tuple = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.train() lowerCAmelCase__ :Optional[Any] = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) lowerCAmelCase__ :List[Any] = model(**lowerCAmelCase__ ).loss loss.backward() def snake_case ( self ): '''simple docstring''' if not self.model_tester.is_training: return for model_class in self.all_model_classes: lowerCAmelCase__ :int = self.model_tester.prepare_config_and_inputs_with_labels() lowerCAmelCase__ :Union[str, Any] = False lowerCAmelCase__ :int = True if ( model_class.__name__ in [*get_values(lowerCAmelCase__ ), *get_values(lowerCAmelCase__ )] or not model_class.supports_gradient_checkpointing ): continue lowerCAmelCase__ :Optional[Any] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.gradient_checkpointing_enable() model.train() lowerCAmelCase__ :Dict = self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ , return_labels=lowerCAmelCase__ ) lowerCAmelCase__ :Optional[Any] = model(**lowerCAmelCase__ ).loss loss.backward() def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ :Optional[int] = model_class(lowerCAmelCase__ ) lowerCAmelCase__ :Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ :List[str] = [*signature.parameters.keys()] lowerCAmelCase__ :List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowerCAmelCase__ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def snake_case ( self ): '''simple docstring''' def check_hidden_states_output(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): lowerCAmelCase__ :Optional[int] = model_class(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() with torch.no_grad(): lowerCAmelCase__ :Union[str, Any] = model(**self._prepare_for_class(lowerCAmelCase__ , lowerCAmelCase__ ) ) lowerCAmelCase__ :List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase__ :Union[str, Any] = self.model_tester.num_stages self.assertEqual(len(lowerCAmelCase__ ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCAmelCase__ :Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ :str = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ :str = True check_hidden_states_output(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__ ) @slow def snake_case ( self ): '''simple docstring''' for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ :List[Any] = ConvNextVaModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def __A () ->Any: """simple docstring""" lowerCAmelCase__ :Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case ( self ): '''simple docstring''' return AutoImageProcessor.from_pretrained('facebook/convnextv2-tiny-1k-224' ) if is_vision_available() else None @slow def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Dict = ConvNextVaForImageClassification.from_pretrained('facebook/convnextv2-tiny-1k-224' ).to(lowerCAmelCase__ ) lowerCAmelCase__ :Tuple = self.default_image_processor lowerCAmelCase__ :Dict = prepare_img() lowerCAmelCase__ :List[str] = preprocessor(images=lowerCAmelCase__ , return_tensors='pt' ).to(lowerCAmelCase__ ) # forward pass with torch.no_grad(): lowerCAmelCase__ :str = model(**lowerCAmelCase__ ) # verify the logits lowerCAmelCase__ :List[str] = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase__ ) lowerCAmelCase__ :int = torch.tensor([0.99_96, 0.19_66, -0.43_86] ).to(lowerCAmelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4 ) )
293
'''simple docstring''' import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _UpperCamelCase : '''simple docstring''' @staticmethod def UpperCamelCase__ ( *lowerCAmelCase__ : Any , **lowerCAmelCase__ : Any ): """simple docstring""" pass @is_pipeline_test @require_vision @require_torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' _A : Optional[int] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def UpperCamelCase__ ( self : str , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Optional[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) __SCREAMING_SNAKE_CASE : Any = [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] return object_detector, examples def UpperCamelCase__ ( self : int , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = object_detector(examples[0] , threshold=0.0 ) __SCREAMING_SNAKE_CASE : Optional[Any] = len(lowerCAmelCase__ ) self.assertGreater(lowerCAmelCase__ , 0 ) self.assertEqual( lowerCAmelCase__ , [ { """score""": ANY(lowerCAmelCase__ ), """label""": ANY(lowerCAmelCase__ ), """box""": {"""xmin""": ANY(lowerCAmelCase__ ), """ymin""": ANY(lowerCAmelCase__ ), """xmax""": ANY(lowerCAmelCase__ ), """ymax""": ANY(lowerCAmelCase__ )}, } for i in range(lowerCAmelCase__ ) ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def UpperCamelCase__ ( self : Optional[int] ): """simple docstring""" pass @require_torch def UpperCamelCase__ ( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) __SCREAMING_SNAKE_CASE : int = object_detector( """./tests/fixtures/tests_samples/COCO/000000039769.png""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=0.64 , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.72_35, """label""": """cat""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.72_18, """label""": """remote""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.71_84, """label""": """couch""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.67_48, """label""": """remote""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_56, """label""": """cat""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_14, """label""": """couch""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.64_56, """label""": """remote""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, {"""score""": 0.6_42, """label""": """remote""", """box""": {"""xmin""": 6_7, """ymin""": 2_7_4, """xmax""": 9_3, """ymax""": 2_9_7}}, {"""score""": 0.64_19, """label""": """cat""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, ] , ) __SCREAMING_SNAKE_CASE : List[Any] = object_detector( [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"""score""": 0.72_35, """label""": """cat""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.72_18, """label""": """remote""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.71_84, """label""": """couch""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.67_48, """label""": """remote""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_56, """label""": """cat""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_14, """label""": """couch""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.64_56, """label""": """remote""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, {"""score""": 0.6_42, """label""": """remote""", """box""": {"""xmin""": 6_7, """ymin""": 2_7_4, """xmax""": 9_3, """ymax""": 2_9_7}}, {"""score""": 0.64_19, """label""": """cat""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, ] ] , ) @require_torch @slow def UpperCamelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = pipeline("""zero-shot-object-detection""" ) __SCREAMING_SNAKE_CASE : List[str] = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, {"""score""": 0.14_74, """label""": """remote""", """box""": {"""xmin""": 3_3_5, """ymin""": 7_4, """xmax""": 3_7_1, """ymax""": 1_8_7}}, {"""score""": 0.12_08, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 6_4_2, """ymax""": 4_7_6}}, ] , ) __SCREAMING_SNAKE_CASE : Dict = object_detector( [ { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, ] , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, {"""score""": 0.14_74, """label""": """remote""", """box""": {"""xmin""": 3_3_5, """ymin""": 7_4, """xmax""": 3_7_1, """ymax""": 1_8_7}}, {"""score""": 0.12_08, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 6_4_2, """ymax""": 4_7_6}}, ], [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, {"""score""": 0.14_74, """label""": """remote""", """box""": {"""xmin""": 3_3_5, """ymin""": 7_4, """xmax""": 3_7_1, """ymax""": 1_8_7}}, {"""score""": 0.12_08, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 6_4_2, """ymax""": 4_7_6}}, ], ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def UpperCamelCase__ ( self : Any ): """simple docstring""" pass @require_torch @slow def UpperCamelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = 0.2 __SCREAMING_SNAKE_CASE : Dict = pipeline("""zero-shot-object-detection""" ) __SCREAMING_SNAKE_CASE : Any = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, ] , ) @require_torch @slow def UpperCamelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = 2 __SCREAMING_SNAKE_CASE : int = pipeline("""zero-shot-object-detection""" ) __SCREAMING_SNAKE_CASE : int = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , top_k=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, ] , )
112
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Dict = { "configuration_lilt": ["LILT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LiltConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : List[str] = [ "LILT_PRETRAINED_MODEL_ARCHIVE_LIST", "LiltForQuestionAnswering", "LiltForSequenceClassification", "LiltForTokenClassification", "LiltModel", "LiltPreTrainedModel", ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys _UpperCAmelCase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
365
from __future__ import annotations class lowercase : def __init__( self , A_ , A_ ) -> Any: """simple docstring""" UpperCamelCase , UpperCamelCase = text, pattern UpperCamelCase , UpperCamelCase = len(A_ ), len(A_ ) def __UpperCamelCase ( self , A_ ) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def __UpperCamelCase ( self , A_ ) -> int: """simple docstring""" for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def __UpperCamelCase ( self ) -> list[int]: """simple docstring""" # searches pattern in text and returns index positions UpperCamelCase = [] for i in range(self.textLen - self.patLen + 1 ): UpperCamelCase = self.mismatch_in_text(A_ ) if mismatch_index == -1: positions.append(A_ ) else: UpperCamelCase = self.match_in_pattern(self.text[mismatch_index] ) UpperCamelCase = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _UpperCAmelCase : Union[str, Any] = "ABAABA" _UpperCAmelCase : Any = "AB" _UpperCAmelCase : Dict = BoyerMooreSearch(text, pattern) _UpperCAmelCase : Optional[int] = bms.bad_character_heuristic() if len(positions) == 0: print("No match found") else: print("Pattern found in following positions: ") print(positions)
110
0
"""simple docstring""" from torch import nn class __snake_case ( nn.Module): def __init__( self : Any , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Union[str, Any] ): """simple docstring""" super().__init__() _lowerCamelCase : Optional[Any] = class_size _lowerCamelCase : Optional[Any] = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) _lowerCamelCase : List[str] = nn.Linear(__lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : int ): """simple docstring""" _lowerCamelCase : List[Any] = self.mlp(__lowerCAmelCase ) return logits
72
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def snake_case_ ( A_ : str, A_ : str, A_ : Optional[str] = None ): '''simple docstring''' if version.parse(hfh.__version__ ).release < version.parse('''0.11.0''' ).release: # old versions of hfh don't url-encode the file path _lowerCamelCase : Optional[Any] = quote(A_ ) return hfh.hf_hub_url(A_, A_, repo_type='''dataset''', revision=A_ )
72
1
'''simple docstring''' import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def lowerCamelCase__ ( __lowerCamelCase : Any , __lowerCamelCase : List[str] , __lowerCamelCase : int=None ): '''simple docstring''' assert torch_layer.weight.shape == weight.shape, f"{torch_layer} layer.weight does not match" _UpperCAmelCase : Union[str, Any] =nn.Parameter(__lowerCamelCase ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"{torch_layer} layer.bias does not match" _UpperCAmelCase : Optional[int] =nn.Parameter(__lowerCamelCase ) def lowerCamelCase__ ( __lowerCamelCase : List[Any] , __lowerCamelCase : List[str] , __lowerCamelCase : List[Any] ): '''simple docstring''' _UpperCAmelCase : Any =np.asarray(weights[0] ) _UpperCAmelCase : Optional[Any] =np.asarray(weights[1] ) _UpperCAmelCase : Optional[int] =np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key , torch.tensor(__lowerCamelCase ).transpose(1 , 2 ).contiguous().view(-1 , __lowerCamelCase ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__lowerCamelCase ).transpose(1 , 2 ).contiguous().view(-1 , __lowerCamelCase ) , ) set_param( torch_layer.output.dense , torch.tensor(__lowerCamelCase ).view(-1 , __lowerCamelCase ).contiguous().transpose(0 , 1 ) , ) def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : int ): '''simple docstring''' _UpperCAmelCase : Any =np.asarray(weights[0] ) _UpperCAmelCase : Dict =np.asarray(weights[1] ) _UpperCAmelCase : List[str] =np.asarray(weights[2] ) _UpperCAmelCase : Optional[int] =np.asarray(weights[3] ) set_param( torch_layer.self_attention.query , torch.tensor(__lowerCamelCase ).transpose(1 , 2 ).contiguous().view(-1 , __lowerCamelCase ) , ) set_param( torch_layer.self_attention.key , torch.tensor(__lowerCamelCase ).transpose(1 , 2 ).contiguous().view(-1 , __lowerCamelCase ) , ) set_param( torch_layer.self_attention.value , torch.tensor(__lowerCamelCase ).transpose(1 , 2 ).contiguous().view(-1 , __lowerCamelCase ) , ) set_param( torch_layer.output.dense , torch.tensor(__lowerCamelCase ).view(-1 , __lowerCamelCase ).contiguous().transpose(0 , 1 ) , ) def lowerCamelCase__ ( __lowerCamelCase : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : Union[str, Any] ): '''simple docstring''' _UpperCAmelCase : Tuple =weights[0][0][0] _UpperCAmelCase : Dict =np.asarray(layer_norm_a[0] ) _UpperCAmelCase : List[str] =np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm , torch.tensor(__lowerCamelCase ) , torch.tensor(__lowerCamelCase ) , ) # lsh weights + output _UpperCAmelCase : Tuple =weights[0][1] if len(__lowerCamelCase ) < 4: set_layer_weights_in_torch_lsh(__lowerCamelCase , torch_block.attention , __lowerCamelCase ) else: set_layer_weights_in_torch_local(__lowerCamelCase , torch_block.attention , __lowerCamelCase ) # intermediate weighs _UpperCAmelCase : str =weights[2][0][1][2] # Chunked Feed Forward if len(__lowerCamelCase ) == 4: _UpperCAmelCase : str =intermediate_weights[2] # layernorm 2 _UpperCAmelCase : Union[str, Any] =np.asarray(intermediate_weights[0][0] ) _UpperCAmelCase : Union[str, Any] =np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm , torch.tensor(__lowerCamelCase ) , torch.tensor(__lowerCamelCase ) , ) # intermediate dense _UpperCAmelCase : Dict =np.asarray(intermediate_weights[1][0] ) _UpperCAmelCase : Dict =np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense , torch.tensor(__lowerCamelCase ).transpose(0 , 1 ).contiguous() , torch.tensor(__lowerCamelCase ) , ) # intermediate out _UpperCAmelCase : int =np.asarray(intermediate_weights[4][0] ) _UpperCAmelCase : List[str] =np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense , torch.tensor(__lowerCamelCase ).transpose(0 , 1 ).contiguous() , torch.tensor(__lowerCamelCase ) , ) def lowerCamelCase__ ( __lowerCamelCase : str , __lowerCamelCase : int , __lowerCamelCase : Optional[Any] ): '''simple docstring''' _UpperCAmelCase : Tuple =torch_model.reformer # word embeds _UpperCAmelCase : str =np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings , torch.tensor(__lowerCamelCase ) , ) if isinstance(weights[3] , __lowerCamelCase ): _UpperCAmelCase : Optional[Any] =torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): _UpperCAmelCase : Union[str, Any] =np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"{position_embeddings[emb_idx]} emb does not match" _UpperCAmelCase : List[Any] =nn.Parameter(torch.tensor(__lowerCamelCase ) ) _UpperCAmelCase : List[Any] =weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __lowerCamelCase ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): _UpperCAmelCase : List[str] =trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # output layer norm _UpperCAmelCase : Union[str, Any] =np.asarray(weights[7][0] ) _UpperCAmelCase : str =np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm , torch.tensor(__lowerCamelCase ) , torch.tensor(__lowerCamelCase ) , ) # output embeddings _UpperCAmelCase : Union[str, Any] =np.asarray(weights[9][0] ) _UpperCAmelCase : List[str] =np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder , torch.tensor(__lowerCamelCase ).transpose(0 , 1 ).contiguous() , torch.tensor(__lowerCamelCase ) , ) def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : Dict , __lowerCamelCase : Dict ): '''simple docstring''' _UpperCAmelCase : Dict =ReformerConfig.from_json_file(__lowerCamelCase ) print(f"Building PyTorch model from configuration: {config}" ) _UpperCAmelCase : Optional[Any] =ReformerModelWithLMHead(__lowerCamelCase ) with open(__lowerCamelCase , 'rb' ) as f: _UpperCAmelCase : Optional[int] =pickle.load(__lowerCamelCase )['weights'] set_model_weights_in_torch(__lowerCamelCase , __lowerCamelCase , config.hidden_size ) # Save pytorch-model print(f"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , __lowerCamelCase ) if __name__ == "__main__": lowercase =argparse.ArgumentParser() # Required parameters parser.add_argument( '--trax_model_pkl_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained Reformer model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) lowercase =parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
242
'''simple docstring''' from string import ascii_uppercase lowercase ={str(ord(c) - 55): c for c in ascii_uppercase} def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : int ): '''simple docstring''' if isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError('int() can\'t convert non-string with explicit base' ) if num < 0: raise ValueError('parameter must be positive int' ) if isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError('\'str\' object cannot be interpreted as an integer' ) if isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError('\'float\' object cannot be interpreted as an integer' ) if base in (0, 1): raise ValueError('base must be >= 2' ) if base > 3_6: raise ValueError('base must be <= 36' ) _UpperCAmelCase : Union[str, Any] ='' _UpperCAmelCase : Optional[int] =0 _UpperCAmelCase : str =0 while div != 1: _UpperCAmelCase , _UpperCAmelCase : int =divmod(__lowerCamelCase , __lowerCamelCase ) if base >= 1_1 and 9 < mod < 3_6: _UpperCAmelCase : str =ALPHABET_VALUES[str(__lowerCamelCase )] else: _UpperCAmelCase : Any =str(__lowerCamelCase ) new_value += actual_value _UpperCAmelCase : Union[str, Any] =num // base _UpperCAmelCase : Dict =div if div == 0: return str(new_value[::-1] ) elif div == 1: new_value += str(__lowerCamelCase ) return str(new_value[::-1] ) return new_value[::-1] if __name__ == "__main__": import doctest doctest.testmod() for base in range(2, 37): for num in range(1000): assert int(decimal_to_any(num, base), base) == num, ( num, base, decimal_to_any(num, base), int(decimal_to_any(num, base), base), )
242
1
"""simple docstring""" import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class _UpperCAmelCase : def __init__( self :Optional[int] , __UpperCamelCase :Tuple=2 , __UpperCamelCase :List[Any]=3 , __UpperCamelCase :Any=64 , __UpperCamelCase :Tuple=None ): A = np.random.default_rng(__UpperCamelCase ) A = length A = rng.normal(size=(length,) ).astype(np.floataa ) A = a * self.x + b + rng.normal(scale=0.1 , size=(length,) ).astype(np.floataa ) def __len__( self :Tuple ): return self.length def __getitem__( self :str , __UpperCamelCase :Optional[int] ): return {"x": self.x[i], "y": self.y[i]} class _UpperCAmelCase ( torch.nn.Module ): def __init__( self :Tuple , __UpperCamelCase :List[str]=0 , __UpperCamelCase :Optional[int]=0 , __UpperCamelCase :Optional[int]=False ): super().__init__() A = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) A = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) A = True def lowerCamelCase ( self :Dict , __UpperCamelCase :Optional[int]=None ): if self.first_batch: print(f"Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}" ) A = False return x * self.a[0] + self.b[0] class _UpperCAmelCase ( torch.nn.Module ): def __init__( self :str , __UpperCamelCase :List[str]=0 , __UpperCamelCase :List[str]=0 , __UpperCamelCase :str=False ): super().__init__() A = torch.nn.Parameter(torch.tensor(__UpperCamelCase ).float() ) A = torch.nn.Parameter(torch.tensor(__UpperCamelCase ).float() ) A = True def lowerCamelCase ( self :int , __UpperCamelCase :Union[str, Any]=None ): if self.first_batch: print(f"Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}" ) A = False return x * self.a + self.b def A__ ( UpperCamelCase , UpperCamelCase = 16 ): from datasets import load_dataset from transformers import AutoTokenizer A = AutoTokenizer.from_pretrained("bert-base-cased" ) A = {"train": "tests/test_samples/MRPC/train.csv", "validation": "tests/test_samples/MRPC/dev.csv"} A = load_dataset("csv" , data_files=UpperCamelCase ) A = datasets["train"].unique("label" ) A = {v: i for i, v in enumerate(UpperCamelCase )} def tokenize_function(UpperCamelCase ): # max_length=None => use the model max length (it's actually the default) A = tokenizer( examples["sentence1"] , examples["sentence2"] , truncation=UpperCamelCase , max_length=UpperCamelCase , padding="max_length" ) if "label" in examples: A = [label_to_id[l] for l in examples["label"]] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset A = datasets.map( UpperCamelCase , batched=UpperCamelCase , remove_columns=["sentence1", "sentence2", "label"] , ) def collate_fn(UpperCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(UpperCamelCase , padding="max_length" , max_length=128 , return_tensors="pt" ) return tokenizer.pad(UpperCamelCase , padding="longest" , return_tensors="pt" ) # Instantiate dataloaders. A = DataLoader(tokenized_datasets["train"] , shuffle=UpperCamelCase , collate_fn=UpperCamelCase , batch_size=2 ) A = DataLoader(tokenized_datasets["validation"] , shuffle=UpperCamelCase , collate_fn=UpperCamelCase , batch_size=1 ) return train_dataloader, eval_dataloader
292
"""simple docstring""" # A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also say that there is no edge that connects # vertices of same set. def A__ ( UpperCamelCase ): A = [False] * len(UpperCamelCase ) A = [-1] * len(UpperCamelCase ) def dfs(UpperCamelCase , UpperCamelCase ): A = True A = c for u in graph[v]: if not visited[u]: dfs(UpperCamelCase , 1 - c ) for i in range(len(UpperCamelCase ) ): if not visited[i]: dfs(UpperCamelCase , 0 ) for i in range(len(UpperCamelCase ) ): for j in graph[i]: if color[i] == color[j]: return False return True # Adjacency list of graph _snake_case : str = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []} print(check_bipartite_dfs(graph))
292
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowercase__ = { """configuration_graphormer""": ["""GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GraphormerConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ = [ """GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """GraphormerForGraphClassification""", """GraphormerModel""", """GraphormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_graphormer import GRAPHORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, GraphormerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_graphormer import ( GRAPHORMER_PRETRAINED_MODEL_ARCHIVE_LIST, GraphormerForGraphClassification, GraphormerModel, GraphormerPreTrainedModel, ) else: import sys lowercase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
353
"""simple docstring""" def __lowerCamelCase ( __UpperCamelCase = 50 ) -> int: """simple docstring""" lowerCAmelCase_ : int = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(F"""{solution() = }""")
161
0
from __future__ import annotations import math def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_UpperCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True _UpperCAmelCase : Any = [num for num in range(3, 10_00_01, 2) if not is_prime(num)] def SCREAMING_SNAKE_CASE ( _UpperCAmelCase ) -> list[int]: if not isinstance(_UpperCAmelCase , _UpperCAmelCase ): raise ValueError('n must be an integer' ) if n <= 0: raise ValueError('n must be >= 0' ) lowerCamelCase__ : int = [] for num in range(len(_UpperCAmelCase ) ): lowerCamelCase__ : Union[str, Any] = 0 while 2 * i * i <= odd_composites[num]: lowerCamelCase__ : Dict = odd_composites[num] - 2 * i * i if is_prime(_UpperCAmelCase ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(_UpperCAmelCase ) == n: return list_nums return [] def SCREAMING_SNAKE_CASE ( ) -> int: return compute_nums(1 )[0] if __name__ == "__main__": print(F"""{solution() = }""")
50
"""simple docstring""" from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _A = 10 def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> int: for i in range(lowerCAmelCase , lowerCAmelCase ): if array[i] == target: return i return -1 def a__ ( lowerCAmelCase , lowerCAmelCase ) -> int: UpperCAmelCase__ : str = 0 UpperCAmelCase__ : Dict = len(lowerCAmelCase ) while left <= right: if right - left < precision: return lin_search(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase__ : Union[str, Any] = (left + right) // 3 + 1 UpperCAmelCase__ : Optional[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: UpperCAmelCase__ : str = one_third - 1 elif array[two_third] < target: UpperCAmelCase__ : Tuple = two_third + 1 else: UpperCAmelCase__ : Any = one_third + 1 UpperCAmelCase__ : str = two_third - 1 else: return -1 def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> int: if left < right: if right - left < precision: return lin_search(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) UpperCAmelCase__ : List[str] = (left + right) // 3 + 1 UpperCAmelCase__ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(lowerCAmelCase , one_third - 1 , lowerCAmelCase , lowerCAmelCase ) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) else: return rec_ternary_search(one_third + 1 , two_third - 1 , lowerCAmelCase , lowerCAmelCase ) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _A = input("""Enter numbers separated by comma:\n""").strip() _A = [int(item.strip()) for item in user_input.split(""",""")] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _A = int(input("""Enter the number to be found in the list:\n""").strip()) _A = ite_ternary_search(collection, target) _A = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f'''Iterative search: {target} found at positions: {resulta}''') print(f'''Recursive search: {target} found at positions: {resulta}''') else: print("""Not found""")
171
0
import baseaa def lowerCAmelCase_ ( __lowerCamelCase ): return baseaa.baaencode(string.encode("utf-8" ) ) def lowerCAmelCase_ ( __lowerCamelCase ): return baseaa.baadecode(__lowerCamelCase ).decode("utf-8" ) if __name__ == "__main__": _snake_case : Dict = "Hello World!" _snake_case : Dict = baseaa_encode(test) print(encoded) _snake_case : Dict = baseaa_decode(encoded) print(decoded)
134
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter _snake_case : Optional[Any] = "Create a default config file for Accelerate with only a few flags set." def lowerCAmelCase_ ( __lowerCamelCase="no" , __lowerCamelCase = default_json_config_file , __lowerCamelCase = False ): __snake_case : int = Path(__lowerCamelCase ) path.parent.mkdir(parents=__lowerCamelCase , exist_ok=__lowerCamelCase ) if path.exists(): print( F'Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.' ) return False __snake_case : Any = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( F'`mixed_precision` should be one of \'no\', \'fp16\', \'bf16\', or \'fp8\'. Received {mixed_precision}' ) __snake_case : Optional[int] = { "compute_environment": "LOCAL_MACHINE", "mixed_precision": mixed_precision, } if torch.cuda.is_available(): __snake_case : Dict = torch.cuda.device_count() __snake_case : Tuple = num_gpus __snake_case : List[str] = False if num_gpus > 1: __snake_case : Optional[int] = "MULTI_GPU" else: __snake_case : Dict = "NO" elif is_xpu_available() and use_xpu: __snake_case : List[str] = torch.xpu.device_count() __snake_case : str = num_xpus __snake_case : int = False if num_xpus > 1: __snake_case : Optional[int] = "MULTI_XPU" else: __snake_case : str = "NO" elif is_npu_available(): __snake_case : Any = torch.npu.device_count() __snake_case : str = num_npus __snake_case : str = False if num_npus > 1: __snake_case : Optional[int] = "MULTI_NPU" else: __snake_case : int = "NO" else: __snake_case : List[Any] = 0 __snake_case : Dict = True __snake_case : Tuple = 1 __snake_case : Tuple = "NO" __snake_case : str = ClusterConfig(**__lowerCamelCase ) config.to_json_file(__lowerCamelCase ) return path def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[Any] = parser.add_parser("default" , parents=__lowerCamelCase , help=__lowerCamelCase , formatter_class=__lowerCamelCase ) parser.add_argument( "--config_file" , default=__lowerCamelCase , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , dest="save_location" , ) parser.add_argument( "--mixed_precision" , choices=["no", "fp16", "bf16"] , type=__lowerCamelCase , help="Whether or not to use mixed precision training. " "Choose between FP16 and BF16 (bfloat16) training. " "BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later." , default="no" , ) parser.set_defaults(func=__lowerCamelCase ) return parser def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(F'accelerate configuration saved at {config_file}' )
134
1
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) _lowerCamelCase : List[Any] = str(bin(lowercase__ ) )[2:] # remove the leading "0b" _lowerCamelCase : Union[str, Any] = str(bin(lowercase__ ) )[2:] _lowerCamelCase : Optional[int] = max(len(lowercase__ ) , len(lowercase__ ) ) return "0b" + "".join( str(int('1' in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowercase__ ) , b_binary.zfill(lowercase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
96
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def snake_case( __magic_name__ , __magic_name__ , __magic_name__ ) -> Optional[Any]: '''simple docstring''' lowercase : int = 1.5 lowercase : int = int(factor * num_class_images ) lowercase : Any = ClipClient( url='''https://knn.laion.ai/knn-service''' , indice_name='''laion_400m''' , num_images=__magic_name__ , aesthetic_weight=0.1 ) os.makedirs(F"""{class_data_dir}/images""" , exist_ok=__magic_name__ ) if len(list(Path(F"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: lowercase : str = client.query(text=__magic_name__ ) if len(__magic_name__ ) >= factor * num_class_images or num_images > 1e4: break else: lowercase : List[str] = int(factor * num_images ) lowercase : List[str] = ClipClient( url='''https://knn.laion.ai/knn-service''' , indice_name='''laion_400m''' , num_images=__magic_name__ , aesthetic_weight=0.1 , ) lowercase : Dict = 0 lowercase : Optional[Any] = 0 lowercase : List[Any] = tqdm(desc='''downloading real regularization images''' , total=__magic_name__ ) with open(F"""{class_data_dir}/caption.txt""" , '''w''' ) as fa, open(F"""{class_data_dir}/urls.txt""" , '''w''' ) as fa, open( F"""{class_data_dir}/images.txt""" , '''w''' ) as fa: while total < num_class_images: lowercase : int = class_images[count] count += 1 try: lowercase : int = requests.get(images['''url'''] ) if img.status_code == 2_00: lowercase : List[Any] = Image.open(BytesIO(img.content ) ) with open(F"""{class_data_dir}/images/{total}.jpg""" , '''wb''' ) as f: f.write(img.content ) fa.write(images['''caption'''] + '''\n''' ) fa.write(images['''url'''] + '''\n''' ) fa.write(F"""{class_data_dir}/images/{total}.jpg""" + '''\n''' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def snake_case( ) -> Optional[int]: '''simple docstring''' lowercase : List[str] = argparse.ArgumentParser('''''' , add_help=__magic_name__ ) parser.add_argument('''--class_prompt''' , help='''text prompt to retrieve images''' , required=__magic_name__ , type=__magic_name__ ) parser.add_argument('''--class_data_dir''' , help='''path to save images''' , required=__magic_name__ , type=__magic_name__ ) parser.add_argument('''--num_class_images''' , help='''number of images to download''' , default=2_00 , type=__magic_name__ ) return parser.parse_args() if __name__ == "__main__": lowerCAmelCase_ = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
308
0
'''simple docstring''' import os def __SCREAMING_SNAKE_CASE ( ): _snake_case = os.path.join(os.path.dirname(_SCREAMING_SNAKE_CASE ) , """num.txt""" ) with open(_SCREAMING_SNAKE_CASE ) as file_hand: return str(sum(int(_SCREAMING_SNAKE_CASE ) for line in file_hand ) )[:10] if __name__ == "__main__": print(solution())
270
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __lowerCAmelCase = { 'configuration_resnet': ['RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ResNetConfig', 'ResNetOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'ResNetForImageClassification', 'ResNetModel', 'ResNetPreTrainedModel', 'ResNetBackbone', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFResNetForImageClassification', 'TFResNetModel', 'TFResNetPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ 'FlaxResNetForImageClassification', 'FlaxResNetModel', 'FlaxResNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['__file__'], _import_structure)
270
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCAmelCase__ ( a , a , a , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ = StableDiffusionInpaintPipeline lowerCAmelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS lowerCAmelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowerCAmelCase__ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess lowerCAmelCase__ = frozenset([] ) def UpperCAmelCase__ ( self : str ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=__SCREAMING_SNAKE_CASE , ) __SCREAMING_SNAKE_CASE = PNDMScheduler(skip_prk_steps=__SCREAMING_SNAKE_CASE ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="""gelu""" , projection_dim=512 , ) __SCREAMING_SNAKE_CASE = CLIPTextModel(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __SCREAMING_SNAKE_CASE = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def UpperCAmelCase__ ( self : Any , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[Any]=0 ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = floats_tensor((1, 3, 32, 32) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ).to(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = image.cpu().permute(0 , 2 , 3 , 1 )[0] __SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(__SCREAMING_SNAKE_CASE ) ).convert("""RGB""" ).resize((64, 64) ) __SCREAMING_SNAKE_CASE = Image.fromarray(np.uinta(image + 4 ) ).convert("""RGB""" ).resize((64, 64) ) if str(__SCREAMING_SNAKE_CASE ).startswith("""mps""" ): __SCREAMING_SNAKE_CASE = torch.manual_seed(__SCREAMING_SNAKE_CASE ) else: __SCREAMING_SNAKE_CASE = torch.Generator(device=__SCREAMING_SNAKE_CASE ).manual_seed(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = { """prompt""": """A painting of a squirrel eating a burger""", """image""": init_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """numpy""", } return inputs def UpperCAmelCase__ ( self : Union[str, Any] ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE = """cpu""" # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE = self.get_dummy_components() __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline(**__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = sd_pipe.to(__SCREAMING_SNAKE_CASE ) sd_pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = self.get_dummy_inputs(__SCREAMING_SNAKE_CASE ) __SCREAMING_SNAKE_CASE = sd_pipe(**__SCREAMING_SNAKE_CASE ).images __SCREAMING_SNAKE_CASE = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCAmelCase__ ( self : Tuple ) -> str: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self : List[Any] ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self : List[str] ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) __SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench.npy""" ) __SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-2-inpainting""" __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline.from_pretrained(__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE = """Face of a yellow cat, high resolution, sitting on a park bench""" __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9E-3 def UpperCAmelCase__ ( self : List[Any] ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) __SCREAMING_SNAKE_CASE = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint""" """/yellow_cat_sitting_on_a_park_bench_fp16.npy""" ) __SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-2-inpainting""" __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline.from_pretrained( __SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , safety_checker=__SCREAMING_SNAKE_CASE , ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE = """Face of a yellow cat, high resolution, sitting on a park bench""" __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __SCREAMING_SNAKE_CASE = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def UpperCAmelCase__ ( self : Tuple ) -> Any: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/sd2-inpaint/init_image.png""" ) __SCREAMING_SNAKE_CASE = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png""" ) __SCREAMING_SNAKE_CASE = """stabilityai/stable-diffusion-2-inpainting""" __SCREAMING_SNAKE_CASE = PNDMScheduler.from_pretrained(__SCREAMING_SNAKE_CASE , subfolder="""scheduler""" ) __SCREAMING_SNAKE_CASE = StableDiffusionInpaintPipeline.from_pretrained( __SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa , ) pipe.to(__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __SCREAMING_SNAKE_CASE = """Face of a yellow cat, high resolution, sitting on a park bench""" __SCREAMING_SNAKE_CASE = torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE = pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , mask_image=__SCREAMING_SNAKE_CASE , generator=__SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type="""np""" , ) __SCREAMING_SNAKE_CASE = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
267
'''simple docstring''' def a__ ( a__ , a__ ): """simple docstring""" _enforce_args(a__ , a__ ) if n == 0: return 0 __SCREAMING_SNAKE_CASE = float("""-inf""" ) for i in range(1 , n + 1 ): __SCREAMING_SNAKE_CASE = max( a__ , prices[i - 1] + naive_cut_rod_recursive(n - i , a__ ) ) return max_revue def a__ ( a__ , a__ ): """simple docstring""" _enforce_args(a__ , a__ ) __SCREAMING_SNAKE_CASE = [float("""-inf""" ) for _ in range(n + 1 )] return _top_down_cut_rod_recursive(a__ , a__ , a__ ) def a__ ( a__ , a__ , a__ ): """simple docstring""" if max_rev[n] >= 0: return max_rev[n] elif n == 0: return 0 else: __SCREAMING_SNAKE_CASE = float("""-inf""" ) for i in range(1 , n + 1 ): __SCREAMING_SNAKE_CASE = max( a__ , prices[i - 1] + _top_down_cut_rod_recursive(n - i , a__ , a__ ) , ) __SCREAMING_SNAKE_CASE = max_revenue return max_rev[n] def a__ ( a__ , a__ ): """simple docstring""" _enforce_args(a__ , a__ ) # length(max_rev) = n + 1, to accommodate for the revenue obtainable from a rod of # length 0. __SCREAMING_SNAKE_CASE = [float("""-inf""" ) for _ in range(n + 1 )] __SCREAMING_SNAKE_CASE = 0 for i in range(1 , n + 1 ): __SCREAMING_SNAKE_CASE = max_rev[i] for j in range(1 , i + 1 ): __SCREAMING_SNAKE_CASE = max(a__ , prices[j - 1] + max_rev[i - j] ) __SCREAMING_SNAKE_CASE = max_revenue_i return max_rev[n] def a__ ( a__ , a__ ): """simple docstring""" if n < 0: __SCREAMING_SNAKE_CASE = F'n must be greater than or equal to 0. Got n = {n}' raise ValueError(a__ ) if n > len(a__ ): __SCREAMING_SNAKE_CASE = ( """Each integral piece of rod must have a corresponding price. """ F'Got n = {n} but length of prices = {len(a__ )}' ) raise ValueError(a__ ) def a__ ( ): """simple docstring""" __SCREAMING_SNAKE_CASE = [6, 10, 12, 15, 20, 23] __SCREAMING_SNAKE_CASE = len(a__ ) # the best revenue comes from cutting the rod into 6 pieces, each # of length 1 resulting in a revenue of 6 * 6 = 36. __SCREAMING_SNAKE_CASE = 36 __SCREAMING_SNAKE_CASE = top_down_cut_rod(a__ , a__ ) __SCREAMING_SNAKE_CASE = bottom_up_cut_rod(a__ , a__ ) __SCREAMING_SNAKE_CASE = naive_cut_rod_recursive(a__ , a__ ) assert expected_max_revenue == max_rev_top_down assert max_rev_top_down == max_rev_bottom_up assert max_rev_bottom_up == max_rev_naive if __name__ == "__main__": main()
267
1
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf __UpperCamelCase : str = logging.get_logger(__name__) @dataclass class __lowerCAmelCase ( lowercase_ ): UpperCamelCase__ = [ "no_inference", "no_cuda", "no_tpu", "no_speed", "no_memory", "no_env_print", "no_multi_process", ] def __init__( self :str , **__magic_name__ :List[str] ): '''simple docstring''' for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: a = deprecated_arg[3:] a = not kwargs.pop(a__ ) logger.warning( F'{deprecated_arg} is depreciated. Please use --no-{positive_arg} or' F' {positive_arg}={kwargs[positive_arg]}' ) a = kwargs.pop("""tpu_name""" , self.tpu_name ) a = kwargs.pop("""device_idx""" , self.device_idx ) a = kwargs.pop("""eager_mode""" , self.eager_mode ) a = kwargs.pop("""use_xla""" , self.use_xla ) super().__init__(**a__ ) UpperCamelCase__ = field( default=lowercase_ , metadata={'''help''': '''Name of TPU'''} , ) UpperCamelCase__ = field( default=0 , metadata={'''help''': '''CPU / GPU device index. Defaults to 0.'''} , ) UpperCamelCase__ = field(default=lowercase_ , metadata={'''help''': '''Benchmark models in eager model.'''} ) UpperCamelCase__ = field( default=lowercase_ , metadata={ '''help''': '''Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.''' } , ) @cached_property def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' requires_backends(self , ["""tf"""] ) a = None if self.tpu: try: if self.tpu_name: a = tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: a = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: a = None return tpu @cached_property def lowerCamelCase__ ( self :int ): '''simple docstring''' requires_backends(self , ["""tf"""] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) a = tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , """GPU""" ) a = tf.distribute.OneDeviceStrategy(device=F'/gpu:{self.device_idx}' ) else: tf.config.set_visible_devices([] , """GPU""" ) # disable GPU a = tf.distribute.OneDeviceStrategy(device=F'/cpu:{self.device_idx}' ) return strategy @property def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' requires_backends(self , ["""tf"""] ) return self._setup_tpu is not None @property def lowerCamelCase__ ( self :int ): '''simple docstring''' requires_backends(self , ["""tf"""] ) return self._setup_strategy @property def lowerCamelCase__ ( self :Optional[int] ): '''simple docstring''' requires_backends(self , ["""tf"""] ) return tf.config.list_physical_devices("""GPU""" ) @property def lowerCamelCase__ ( self :List[Any] ): '''simple docstring''' requires_backends(self , ["""tf"""] ) if self.cuda: return len(self.gpu_list ) return 0 @property def lowerCamelCase__ ( self :List[str] ): '''simple docstring''' return self.n_gpu > 0
354
from __future__ import annotations from typing import Generic, TypeVar __UpperCamelCase : Union[str, Any] = TypeVar("T") class __lowerCAmelCase ( Generic[T] ): def __init__( self :Tuple , __magic_name__ :T ): '''simple docstring''' a = data a = self a = 0 class __lowerCAmelCase ( Generic[T] ): def __init__( self :Tuple ): '''simple docstring''' a = {} def lowerCamelCase__ ( self :Optional[int] , __magic_name__ :T ): '''simple docstring''' a = DisjointSetTreeNode(__magic_name__ ) def lowerCamelCase__ ( self :List[Any] , __magic_name__ :T ): '''simple docstring''' a = self.map[data] if elem_ref != elem_ref.parent: a = self.find_set(elem_ref.parent.data ) return elem_ref.parent def lowerCamelCase__ ( self :List[Any] , __magic_name__ :DisjointSetTreeNode[T] , __magic_name__ :DisjointSetTreeNode[T] ): '''simple docstring''' if nodea.rank > nodea.rank: a = nodea else: a = nodea if nodea.rank == nodea.rank: nodea.rank += 1 def lowerCamelCase__ ( self :Optional[int] , __magic_name__ :T , __magic_name__ :T ): '''simple docstring''' self.link(self.find_set(__magic_name__ ) , self.find_set(__magic_name__ ) ) class __lowerCAmelCase ( Generic[T] ): def __init__( self :Union[str, Any] ): '''simple docstring''' a = {} def lowerCamelCase__ ( self :Union[str, Any] , __magic_name__ :T ): '''simple docstring''' if node not in self.connections: a = {} def lowerCamelCase__ ( self :Any , __magic_name__ :T , __magic_name__ :T , __magic_name__ :int ): '''simple docstring''' self.add_node(__magic_name__ ) self.add_node(__magic_name__ ) a = weight a = weight def lowerCamelCase__ ( self :int ): '''simple docstring''' a = [] a = set() for start in self.connections: for end in self.connections[start]: if (start, end) not in seen: seen.add((end, start) ) edges.append((start, end, self.connections[start][end]) ) edges.sort(key=lambda __magic_name__ : x[2] ) # creating the disjoint set a = DisjointSetTree[T]() for node in self.connections: disjoint_set.make_set(__magic_name__ ) # MST generation a = 0 a = 0 a = GraphUndirectedWeighted[T]() while num_edges < len(self.connections ) - 1: a , a , a = edges[index] index += 1 a = disjoint_set.find_set(__magic_name__ ) a = disjoint_set.find_set(__magic_name__ ) if parent_u != parent_v: num_edges += 1 graph.add_edge(__magic_name__ , __magic_name__ , __magic_name__ ) disjoint_set.union(__magic_name__ , __magic_name__ ) return graph
347
0
'''simple docstring''' import math def __lowerCamelCase ( lowerCAmelCase_ ) -> int: if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _a : Tuple = f"""Input value of [number={number}] must be an integer""" raise TypeError(lowerCAmelCase_ ) if number < 1: _a : List[Any] = f"""Input value of [number={number}] must be > 0""" raise ValueError(lowerCAmelCase_ ) elif number == 1: return 3 elif number == 2: return 5 else: _a : List[Any] = int(math.log(number // 3 , 2 ) ) + 2 _a : str = [3, 5] _a : Optional[int] = 2 _a : Union[str, Any] = 3 for block in range(1 , lowerCAmelCase_ ): for _ in range(lowerCAmelCase_ ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(11): __lowerCAmelCase = 0 try: __lowerCAmelCase = proth(number) except ValueError: print(f"""ValueError: there is no {number}th Proth number""") continue print(f"""The {number}th Proth number: {value}""")
89
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _lowerCamelCase : Tuple = { "configuration_encodec": [ "ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP", "EncodecConfig", ], "feature_extraction_encodec": ["EncodecFeatureExtractor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCamelCase : Optional[Any] = [ "ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST", "EncodecModel", "EncodecPreTrainedModel", ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys _lowerCamelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
336
0
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class UpperCamelCase ( lowercase , lowercase , unittest.TestCase ): UpperCAmelCase : Tuple = IFInpaintingPipeline UpperCAmelCase : Optional[Any] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} UpperCAmelCase : Dict = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS UpperCAmelCase : Dict = PipelineTesterMixin.required_optional_params - {"""latents"""} def _lowercase (self : List[str]) -> Optional[Any]: return self._get_dummy_components() def _lowercase (self : Tuple , _A : Union[str, Any] , _A : int=0) -> Dict: if str(_A).startswith('mps'): __snake_case : Any = torch.manual_seed(_A) else: __snake_case : Any = torch.Generator(device=_A).manual_seed(_A) __snake_case : Optional[Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A)).to(_A) __snake_case : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(_A)).to(_A) __snake_case : str = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'mask_image': mask_image, 'generator': generator, 'num_inference_steps': 2, 'output_type': 'numpy', } return inputs @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def _lowercase (self : Optional[int]) -> Dict: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3) def _lowercase (self : Any) -> Optional[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != 'cuda' , reason='float16 requires CUDA') def _lowercase (self : int) -> Optional[int]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1) def _lowercase (self : Dict) -> List[Any]: self._test_attention_slicing_forward_pass(expected_max_diff=1E-2) def _lowercase (self : str) -> int: self._test_save_load_local() def _lowercase (self : List[Any]) -> List[str]: self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
95
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer _a : List[Any]= logging.get_logger(__name__) _a : Any= {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} _a : int= { "vocab_file": { "junnyu/roformer_chinese_small": "https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt", "junnyu/roformer_chinese_base": "https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt", "junnyu/roformer_chinese_char_small": ( "https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt" ), "junnyu/roformer_chinese_char_base": ( "https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt" ), "junnyu/roformer_small_discriminator": ( "https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt" ), "junnyu/roformer_small_generator": ( "https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt" ), } } _a : Optional[Any]= { "junnyu/roformer_chinese_small": 1_536, "junnyu/roformer_chinese_base": 1_536, "junnyu/roformer_chinese_char_small": 512, "junnyu/roformer_chinese_char_base": 512, "junnyu/roformer_small_discriminator": 128, "junnyu/roformer_small_generator": 128, } _a : str= { "junnyu/roformer_chinese_small": {"do_lower_case": True}, "junnyu/roformer_chinese_base": {"do_lower_case": True}, "junnyu/roformer_chinese_char_small": {"do_lower_case": True}, "junnyu/roformer_chinese_char_base": {"do_lower_case": True}, "junnyu/roformer_small_discriminator": {"do_lower_case": True}, "junnyu/roformer_small_generator": {"do_lower_case": True}, } class UpperCamelCase ( lowercase ): UpperCAmelCase : Union[str, Any] = VOCAB_FILES_NAMES UpperCAmelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : int = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase : Dict = RoFormerTokenizer def __init__(self : List[Any] , _A : Any=None , _A : int=None , _A : Dict=True , _A : List[Any]="[UNK]" , _A : Tuple="[SEP]" , _A : List[Any]="[PAD]" , _A : str="[CLS]" , _A : int="[MASK]" , _A : Optional[int]=True , _A : List[str]=None , **_A : int , ) -> Dict: super().__init__( _A , tokenizer_file=_A , do_lower_case=_A , unk_token=_A , sep_token=_A , pad_token=_A , cls_token=_A , mask_token=_A , tokenize_chinese_chars=_A , strip_accents=_A , **_A , ) __snake_case : List[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( pre_tok_state.get('lowercase' , _A) != do_lower_case or pre_tok_state.get('strip_accents' , _A) != strip_accents ): __snake_case : Union[str, Any] = getattr(_A , pre_tok_state.pop('type')) __snake_case : Union[str, Any] = do_lower_case __snake_case : str = strip_accents __snake_case : Optional[int] = pre_tok_class(**_A) __snake_case : int = do_lower_case def __getstate__(self : Optional[Any]) -> Dict: __snake_case : Optional[int] = self.__dict__.copy() __snake_case : int = BertPreTokenizer() return state def __setstate__(self : Optional[Any] , _A : Optional[Any]) -> Dict: __snake_case : List[str] = d __snake_case : str = self.__dict__['_tokenizer'].get_vocab() __snake_case : int = PreTokenizer.custom(JiebaPreTokenizer(_A)) def _lowercase (self : int , _A : Tuple , _A : Any=None) -> str: __snake_case : Dict = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase (self : List[str] , _A : List[int] , _A : Optional[List[int]] = None) -> List[int]: __snake_case : Tuple = [self.sep_token_id] __snake_case : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def _lowercase (self : List[Any] , _A : str , _A : Optional[str] = None) -> Tuple[str]: __snake_case : List[Any] = self._tokenizer.model.save(_A , name=_A) return tuple(_A) def _lowercase (self : int , _A : Optional[int] , _A : Tuple=None , _A : Tuple=None , _A : Dict=False , **_A : Optional[int] , ) -> Optional[Any]: __snake_case : Optional[Any] = BertPreTokenizer() return super().save_pretrained(_A , _A , _A , _A , **_A)
95
1
from __future__ import annotations import queue class UpperCAmelCase : def __init__(self : Union[str, Any] , snake_case__ : int ) -> Optional[int]: '''simple docstring''' snake_case : int = data snake_case : Optional[Any] = None snake_case : Optional[Any] = None def UpperCamelCase ( ): print("\n********Press N to stop entering at any point of time********\n" ) snake_case : Union[str, Any] = input("Enter the value of the root node: " ).strip().lower() snake_case : queue.Queue = queue.Queue() snake_case : Union[str, Any] = TreeNode(int(__lowerCamelCase ) ) q.put(__lowerCamelCase ) while not q.empty(): snake_case : Optional[Any] = q.get() snake_case : Optional[int] = f"""Enter the left node of {node_found.data}: """ snake_case : Any = input(__lowerCamelCase ).strip().lower() or "n" if check == "n": return tree_node snake_case : int = TreeNode(int(__lowerCamelCase ) ) snake_case : Tuple = left_node q.put(__lowerCamelCase ) snake_case : Dict = f"""Enter the right node of {node_found.data}: """ snake_case : str = input(__lowerCamelCase ).strip().lower() or "n" if check == "n": return tree_node snake_case : Any = TreeNode(int(__lowerCamelCase ) ) snake_case : List[Any] = right_node q.put(__lowerCamelCase ) raise def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return print(node.data , end="," ) pre_order(node.left ) pre_order(node.right ) def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return in_order(node.left ) print(node.data , end="," ) in_order(node.right ) def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end="," ) def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case : queue.Queue = queue.Queue() q.put(__lowerCamelCase ) while not q.empty(): snake_case : int = q.get() print(node_dequeued.data , end="," ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case : queue.Queue = queue.Queue() q.put(__lowerCamelCase ) while not q.empty(): snake_case : Dict = [] while not q.empty(): snake_case : Optional[Any] = q.get() print(node_dequeued.data , end="," ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(__lowerCamelCase ) def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case : list[TreeNode] = [] snake_case : Tuple = node while n or stack: while n: # start from root node, find its left child print(n.data , end="," ) stack.append(__lowerCamelCase ) snake_case : Dict = n.left # end of while means current node doesn't have left child snake_case : Optional[int] = stack.pop() # start to traverse its right child snake_case : List[Any] = n.right def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case : list[TreeNode] = [] snake_case : Dict = node while n or stack: while n: stack.append(__lowerCamelCase ) snake_case : List[Any] = n.left snake_case : List[Any] = stack.pop() print(n.data , end="," ) snake_case : str = n.right def UpperCamelCase ( __lowerCamelCase : TreeNode ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or not node: return snake_case , snake_case : List[Any] = [], [] snake_case : Any = node stacka.append(__lowerCamelCase ) while stacka: # to find the reversed order of post order, store it in stack2 snake_case : Optional[Any] = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(__lowerCamelCase ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end="," ) def UpperCamelCase ( __lowerCamelCase : str = "" , __lowerCamelCase : Tuple=50 , __lowerCamelCase : int="*" ): if not s: return "\n" + width * char snake_case , snake_case : Optional[Any] = divmod(width - len(__lowerCamelCase ) - 2 , 2 ) return f"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt("""Binary Tree Traversals""")) __lowerCamelCase = build_tree() print(prompt("""Pre Order Traversal""")) pre_order(node) print(prompt() + """\n""") print(prompt("""In Order Traversal""")) in_order(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal""")) post_order(node) print(prompt() + """\n""") print(prompt("""Level Order Traversal""")) level_order(node) print(prompt() + """\n""") print(prompt("""Actual Level Order Traversal""")) level_order_actual(node) print("""*""" * 50 + """\n""") print(prompt("""Pre Order Traversal - Iteration Version""")) pre_order_iter(node) print(prompt() + """\n""") print(prompt("""In Order Traversal - Iteration Version""")) in_order_iter(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal - Iteration Version""")) post_order_iter(node) print(prompt())
59
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCamelCase = { """configuration_biogpt""": ["""BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BioGptConfig"""], """tokenization_biogpt""": ["""BioGptTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ """BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BioGptForCausalLM""", """BioGptForTokenClassification""", """BioGptForSequenceClassification""", """BioGptModel""", """BioGptPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys __lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
59
1
"""simple docstring""" import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( ConditionalDetrConfig, ConditionalDetrForObjectDetection, ConditionalDetrForSegmentation, ConditionalDetrImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) __UpperCAmelCase = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", F"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", F"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.weight""", F"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear1.bias""", F"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.weight""", F"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.linear2.bias""", F"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.encoder.layers.{i}.norm1.weight""", F"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.encoder.layers.{i}.norm1.bias""", F"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.weight""", F"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.encoder.layers.{i}.norm2.bias""", F"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", F"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", F"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.cross_attn.out_proj.weight""", F"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( F"""transformer.decoder.layers.{i}.cross_attn.out_proj.bias""", F"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.weight""", F"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear1.bias""", F"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.weight""", F"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.linear2.bias""", F"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm1.weight""", F"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm1.bias""", F"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.weight""", F"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.norm2.bias""", F"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.weight""", F"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((F"""transformer.decoder.layers.{i}.norm3.bias""", F"""decoder.layers.{i}.final_layer_norm.bias""")) # q, k, v projections in self/cross-attention in decoder for conditional DETR rename_keys.append( (F"""transformer.decoder.layers.{i}.sa_qcontent_proj.weight""", F"""decoder.layers.{i}.sa_qcontent_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.sa_kcontent_proj.weight""", F"""decoder.layers.{i}.sa_kcontent_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.sa_qpos_proj.weight""", F"""decoder.layers.{i}.sa_qpos_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.sa_kpos_proj.weight""", F"""decoder.layers.{i}.sa_kpos_proj.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.sa_v_proj.weight""", F"""decoder.layers.{i}.sa_v_proj.weight""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.ca_qcontent_proj.weight""", F"""decoder.layers.{i}.ca_qcontent_proj.weight""") ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.weight", f"decoder.layers.{i}.ca_qpos_proj.weight")) rename_keys.append( (F"""transformer.decoder.layers.{i}.ca_kcontent_proj.weight""", F"""decoder.layers.{i}.ca_kcontent_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.ca_kpos_proj.weight""", F"""decoder.layers.{i}.ca_kpos_proj.weight""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.ca_v_proj.weight""", F"""decoder.layers.{i}.ca_v_proj.weight""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.ca_qpos_sine_proj.weight""", F"""decoder.layers.{i}.ca_qpos_sine_proj.weight""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.sa_qcontent_proj.bias""", F"""decoder.layers.{i}.sa_qcontent_proj.bias""") ) rename_keys.append( (F"""transformer.decoder.layers.{i}.sa_kcontent_proj.bias""", F"""decoder.layers.{i}.sa_kcontent_proj.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.sa_qpos_proj.bias""", F"""decoder.layers.{i}.sa_qpos_proj.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.sa_kpos_proj.bias""", F"""decoder.layers.{i}.sa_kpos_proj.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.sa_v_proj.bias""", F"""decoder.layers.{i}.sa_v_proj.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.ca_qcontent_proj.bias""", F"""decoder.layers.{i}.ca_qcontent_proj.bias""") ) # rename_keys.append((f"transformer.decoder.layers.{i}.ca_qpos_proj.bias", f"decoder.layers.{i}.ca_qpos_proj.bias")) rename_keys.append( (F"""transformer.decoder.layers.{i}.ca_kcontent_proj.bias""", F"""decoder.layers.{i}.ca_kcontent_proj.bias""") ) rename_keys.append((F"""transformer.decoder.layers.{i}.ca_kpos_proj.bias""", F"""decoder.layers.{i}.ca_kpos_proj.bias""")) rename_keys.append((F"""transformer.decoder.layers.{i}.ca_v_proj.bias""", F"""decoder.layers.{i}.ca_v_proj.bias""")) rename_keys.append( (F"""transformer.decoder.layers.{i}.ca_qpos_sine_proj.bias""", F"""decoder.layers.{i}.ca_qpos_sine_proj.bias""") ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads # for conditional DETR, also convert reference point head and query scale MLP rename_keys.extend( [ ('''input_proj.weight''', '''input_projection.weight'''), ('''input_proj.bias''', '''input_projection.bias'''), ('''query_embed.weight''', '''query_position_embeddings.weight'''), ('''transformer.decoder.norm.weight''', '''decoder.layernorm.weight'''), ('''transformer.decoder.norm.bias''', '''decoder.layernorm.bias'''), ('''class_embed.weight''', '''class_labels_classifier.weight'''), ('''class_embed.bias''', '''class_labels_classifier.bias'''), ('''bbox_embed.layers.0.weight''', '''bbox_predictor.layers.0.weight'''), ('''bbox_embed.layers.0.bias''', '''bbox_predictor.layers.0.bias'''), ('''bbox_embed.layers.1.weight''', '''bbox_predictor.layers.1.weight'''), ('''bbox_embed.layers.1.bias''', '''bbox_predictor.layers.1.bias'''), ('''bbox_embed.layers.2.weight''', '''bbox_predictor.layers.2.weight'''), ('''bbox_embed.layers.2.bias''', '''bbox_predictor.layers.2.bias'''), ('''transformer.decoder.ref_point_head.layers.0.weight''', '''decoder.ref_point_head.layers.0.weight'''), ('''transformer.decoder.ref_point_head.layers.0.bias''', '''decoder.ref_point_head.layers.0.bias'''), ('''transformer.decoder.ref_point_head.layers.1.weight''', '''decoder.ref_point_head.layers.1.weight'''), ('''transformer.decoder.ref_point_head.layers.1.bias''', '''decoder.ref_point_head.layers.1.bias'''), ('''transformer.decoder.query_scale.layers.0.weight''', '''decoder.query_scale.layers.0.weight'''), ('''transformer.decoder.query_scale.layers.0.bias''', '''decoder.query_scale.layers.0.bias'''), ('''transformer.decoder.query_scale.layers.1.weight''', '''decoder.query_scale.layers.1.weight'''), ('''transformer.decoder.query_scale.layers.1.bias''', '''decoder.query_scale.layers.1.bias'''), ('''transformer.decoder.layers.0.ca_qpos_proj.weight''', '''decoder.layers.0.ca_qpos_proj.weight'''), ('''transformer.decoder.layers.0.ca_qpos_proj.bias''', '''decoder.layers.0.ca_qpos_proj.bias'''), ] ) def snake_case_ ( A_ : str, A_ : Tuple, A_ : Union[str, Any] ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = state_dict.pop(A_ ) _lowerCamelCase : Union[str, Any] = val def snake_case_ ( A_ : Any ): '''simple docstring''' _lowerCamelCase : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _lowerCamelCase : List[Any] = key.replace('''backbone.0.body''', '''backbone.conv_encoder.model''' ) _lowerCamelCase : int = value else: _lowerCamelCase : List[str] = value return new_state_dict def snake_case_ ( A_ : Optional[int], A_ : List[str]=False ): '''simple docstring''' _lowerCamelCase : Any = '''''' if is_panoptic: _lowerCamelCase : Optional[Any] = '''conditional_detr.''' # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _lowerCamelCase : Optional[int] = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight''' ) _lowerCamelCase : Dict = state_dict.pop(F'''{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase : str = in_proj_weight[:2_56, :] _lowerCamelCase : int = in_proj_bias[:2_56] _lowerCamelCase : str = in_proj_weight[2_56:5_12, :] _lowerCamelCase : Optional[Any] = in_proj_bias[2_56:5_12] _lowerCamelCase : List[Any] = in_proj_weight[-2_56:, :] _lowerCamelCase : List[str] = in_proj_bias[-2_56:] def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase : Any = Image.open(requests.get(A_, stream=A_ ).raw ) return im @torch.no_grad() def snake_case_ ( A_ : Optional[Any], A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : Optional[Any] = ConditionalDetrConfig() # set backbone and dilation attributes if "resnet101" in model_name: _lowerCamelCase : Union[str, Any] = '''resnet101''' if "dc5" in model_name: _lowerCamelCase : Optional[int] = True _lowerCamelCase : Tuple = '''panoptic''' in model_name if is_panoptic: _lowerCamelCase : Optional[int] = 2_50 else: _lowerCamelCase : int = 91 _lowerCamelCase : List[str] = '''huggingface/label-files''' _lowerCamelCase : Any = '''coco-detection-id2label.json''' _lowerCamelCase : Optional[int] = json.load(open(hf_hub_download(A_, A_, repo_type='''dataset''' ), '''r''' ) ) _lowerCamelCase : List[str] = {int(A_ ): v for k, v in idalabel.items()} _lowerCamelCase : List[str] = idalabel _lowerCamelCase : str = {v: k for k, v in idalabel.items()} # load image processor _lowerCamelCase : int = '''coco_panoptic''' if is_panoptic else '''coco_detection''' _lowerCamelCase : Any = ConditionalDetrImageProcessor(format=A_ ) # prepare image _lowerCamelCase : Optional[int] = prepare_img() _lowerCamelCase : str = image_processor(images=A_, return_tensors='''pt''' ) _lowerCamelCase : Union[str, Any] = encoding['''pixel_values'''] logger.info(F'''Converting model {model_name}...''' ) # load original model from torch hub _lowerCamelCase : int = torch.hub.load('''DeppMeng/ConditionalDETR''', A_, pretrained=A_ ).eval() _lowerCamelCase : Tuple = conditional_detr.state_dict() # rename keys for src, dest in rename_keys: if is_panoptic: _lowerCamelCase : Optional[Any] = '''conditional_detr.''' + src rename_key(A_, A_, A_ ) _lowerCamelCase : Dict = rename_backbone_keys(A_ ) # query, key and value matrices need special treatment read_in_q_k_v(A_, is_panoptic=A_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _lowerCamelCase : Optional[int] = '''conditional_detr.model.''' if is_panoptic else '''model.''' for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith('''conditional_detr''' ) and not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ) ): _lowerCamelCase : List[Any] = state_dict.pop(A_ ) _lowerCamelCase : int = val elif "class_labels_classifier" in key or "bbox_predictor" in key: _lowerCamelCase : List[str] = state_dict.pop(A_ ) _lowerCamelCase : Optional[Any] = val elif key.startswith('''bbox_attention''' ) or key.startswith('''mask_head''' ): continue else: _lowerCamelCase : Optional[Any] = state_dict.pop(A_ ) _lowerCamelCase : Any = val else: if not key.startswith('''class_labels_classifier''' ) and not key.startswith('''bbox_predictor''' ): _lowerCamelCase : int = state_dict.pop(A_ ) _lowerCamelCase : str = val # finally, create HuggingFace model and load state dict _lowerCamelCase : Dict = ConditionalDetrForSegmentation(A_ ) if is_panoptic else ConditionalDetrForObjectDetection(A_ ) model.load_state_dict(A_ ) model.eval() model.push_to_hub(repo_id=A_, organization='''DepuMeng''', commit_message='''Add model''' ) # verify our conversion _lowerCamelCase : Dict = conditional_detr(A_ ) _lowerCamelCase : Optional[int] = model(A_ ) assert torch.allclose(outputs.logits, original_outputs['''pred_logits'''], atol=1E-4 ) assert torch.allclose(outputs.pred_boxes, original_outputs['''pred_boxes'''], atol=1E-4 ) if is_panoptic: assert torch.allclose(outputs.pred_masks, original_outputs['''pred_masks'''], atol=1E-4 ) # Save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(A_ ).mkdir(exist_ok=A_ ) model.save_pretrained(A_ ) image_processor.save_pretrained(A_ ) if __name__ == "__main__": __UpperCAmelCase = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''conditional_detr_resnet50''', type=str, help='''Name of the CONDITIONAL_DETR model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) __UpperCAmelCase = parser.parse_args() convert_conditional_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path)
365
"""simple docstring""" from torch import nn class __snake_case ( nn.Module): def __init__( self : Optional[Any] , __lowerCAmelCase : Dict , __lowerCAmelCase : str ): """simple docstring""" super().__init__() _lowerCamelCase : List[str] = class_size _lowerCamelCase : List[str] = embed_size # self.mlp1 = nn.Linear(embed_size, embed_size) # self.mlp2 = (nn.Linear(embed_size, class_size)) _lowerCamelCase : Optional[int] = nn.Linear(__lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : Optional[Any] = self.mlp(__lowerCAmelCase ) return logits
175
0
"""simple docstring""" # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar __magic_name__ = TypeVar("T") class SCREAMING_SNAKE_CASE_ ( Generic[T] ): """simple docstring""" def __init__( self , lowerCAmelCase__ = True): __SCREAMING_SNAKE_CASE = {} # dictionary of lists __SCREAMING_SNAKE_CASE = directed def snake_case_ ( self , lowerCAmelCase__ , lowerCAmelCase__): if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCAmelCase__) self.adj_list[destination_vertex].append(lowerCAmelCase__) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: __SCREAMING_SNAKE_CASE = [destination_vertex] __SCREAMING_SNAKE_CASE = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCAmelCase__) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(lowerCAmelCase__) __SCREAMING_SNAKE_CASE = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: __SCREAMING_SNAKE_CASE = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: __SCREAMING_SNAKE_CASE = [destination_vertex] __SCREAMING_SNAKE_CASE = [] return self def __repr__( self): return pformat(self.adj_list)
100
from pickle import UnpicklingError import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict from ..utils import logging lowerCAmelCase = logging.get_logger(__name__) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" try: with open(SCREAMING_SNAKE_CASE , '''rb''' ) as flax_state_f: lowercase__ = from_bytes(SCREAMING_SNAKE_CASE , flax_state_f.read() ) except UnpicklingError as e: try: with open(SCREAMING_SNAKE_CASE ) as f: if f.read().startswith('''version''' ): raise OSError( '''You seem to have cloned a repository without having git-lfs installed. Please''' ''' install git-lfs and run `git lfs install` followed by `git lfs pull` in the''' ''' folder you cloned.''' ) else: raise ValueError from e except (UnicodeDecodeError, ValueError): raise EnvironmentError(f'Unable to convert {model_file} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( '''Loading Flax weights in PyTorch requires both PyTorch and Flax to be installed. Please see''' ''' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation''' ''' instructions.''' ) raise # check if we have bf16 weights lowercase__ = flatten_dict(jax.tree_util.tree_map(lambda SCREAMING_SNAKE_CASE : x.dtype == jnp.bfloataa , SCREAMING_SNAKE_CASE ) ).values() if any(SCREAMING_SNAKE_CASE ): # convert all weights to fp32 if they are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( '''Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ''' '''before loading those in PyTorch model.''' ) lowercase__ = jax.tree_util.tree_map( lambda SCREAMING_SNAKE_CASE : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , SCREAMING_SNAKE_CASE ) lowercase__ = '''''' lowercase__ = flatten_dict(SCREAMING_SNAKE_CASE , sep='''.''' ) lowercase__ = pt_model.state_dict() # keep track of unexpected & missing keys lowercase__ = [] lowercase__ = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): lowercase__ = flax_key_tuple.split('''.''' ) if flax_key_tuple_array[-1] == "kernel" and flax_tensor.ndim == 4: lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] lowercase__ = jnp.transpose(SCREAMING_SNAKE_CASE , (3, 2, 0, 1) ) elif flax_key_tuple_array[-1] == "kernel": lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] lowercase__ = flax_tensor.T elif flax_key_tuple_array[-1] == "scale": lowercase__ = flax_key_tuple_array[:-1] + ['''weight'''] if "time_embedding" not in flax_key_tuple_array: for i, flax_key_tuple_string in enumerate(SCREAMING_SNAKE_CASE ): lowercase__ = ( flax_key_tuple_string.replace('''_0''' , '''.0''' ) .replace('''_1''' , '''.1''' ) .replace('''_2''' , '''.2''' ) .replace('''_3''' , '''.3''' ) .replace('''_4''' , '''.4''' ) .replace('''_5''' , '''.5''' ) .replace('''_6''' , '''.6''' ) .replace('''_7''' , '''.7''' ) .replace('''_8''' , '''.8''' ) .replace('''_9''' , '''.9''' ) ) lowercase__ = '''.'''.join(SCREAMING_SNAKE_CASE ) if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' f'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict lowercase__ = np.asarray(SCREAMING_SNAKE_CASE ) if not isinstance(SCREAMING_SNAKE_CASE , np.ndarray ) else flax_tensor lowercase__ = torch.from_numpy(SCREAMING_SNAKE_CASE ) # remove from missing keys missing_keys.remove(SCREAMING_SNAKE_CASE ) else: # weight is not expected by PyTorch model unexpected_keys.append(SCREAMING_SNAKE_CASE ) pt_model.load_state_dict(SCREAMING_SNAKE_CASE ) # re-transform missing_keys to list lowercase__ = list(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0: logger.warning( '''Some weights of the Flax model were not used when initializing the PyTorch model''' f' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' f' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' ''' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This''' f' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' ''' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a''' ''' FlaxBertForSequenceClassification model).''' ) if len(SCREAMING_SNAKE_CASE ) > 0: logger.warning( f'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' f' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' ''' use it for predictions and inference.''' ) return pt_model
110
0
import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _lowercase: Dict = logging.get_logger(__name__) class _lowercase ( lowerCAmelCase ): """simple docstring""" __A = ["input_ids", "attention_mask"] def __init__(self , lowerCamelCase_="</s>" , lowerCamelCase_="<unk>" , lowerCamelCase_="<pad>" , lowerCamelCase_=125 , lowerCamelCase_=None , **lowerCamelCase_ , ): """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: a = [F'''<extra_id_{i}>''' for i in range(lowerCamelCase_ )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens a = len(set(filter(lambda lowerCamelCase_ : bool("extra_id" in str(lowerCamelCase_ ) ) , lowerCamelCase_ ) ) ) if extra_tokens != extra_ids: raise ValueError( F'''Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are''' " provided to ByT5Tokenizer. In this case the additional_special_tokens must include the" " extra_ids tokens" ) a = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else pad_token a = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else eos_token a = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else unk_token super().__init__( eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , extra_ids=lowerCamelCase_ , additional_special_tokens=lowerCamelCase_ , **lowerCamelCase_ , ) a = extra_ids a = 2**8 # utf is 8 bits # define special tokens dict a = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } a = len(self.special_tokens_encoder ) a = len(lowerCamelCase_ ) for i, token in enumerate(lowerCamelCase_ ): a = self.vocab_size + i - n a = {v: k for k, v in self.special_tokens_encoder.items()} @property def UpperCamelCase_ (self ): """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_ ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowerCamelCase_ )) + [1] return ([0] * len(lowerCamelCase_ )) + [1] + ([0] * len(lowerCamelCase_ )) + [1] def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" if len(lowerCamelCase_ ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F'''This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated''' " eos tokens being added." ) return token_ids else: return token_ids + [self.eos_token_id] def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ = None ): """simple docstring""" a = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ = None ): """simple docstring""" a = self._add_eos_if_not_present(lowerCamelCase_ ) if token_ids_a is None: return token_ids_a else: a = self._add_eos_if_not_present(lowerCamelCase_ ) return token_ids_a + token_ids_a def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" a = [chr(lowerCamelCase_ ) for i in text.encode("utf-8" )] return tokens def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" if token in self.special_tokens_encoder: a = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: a = self.added_tokens_encoder[token] elif len(lowerCamelCase_ ) != 1: a = self.unk_token_id else: a = ord(lowerCamelCase_ ) + self._num_special_tokens return token_id def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" if index in self.special_tokens_decoder: a = self.special_tokens_decoder[index] else: a = chr(index - self._num_special_tokens ) return token def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" a = B"" for token in tokens: if token in self.special_tokens_decoder: a = self.special_tokens_decoder[token].encode("utf-8" ) elif token in self.added_tokens_decoder: a = self.special_tokens_decoder[token].encode("utf-8" ) elif token in self.special_tokens_encoder: a = token.encode("utf-8" ) elif token in self.added_tokens_encoder: a = token.encode("utf-8" ) else: a = bytes([ord(lowerCamelCase_ )] ) bstring += tok_string a = bstring.decode("utf-8" , errors="ignore" ) return string def UpperCamelCase_ (self , lowerCamelCase_ , lowerCamelCase_ = None ): """simple docstring""" return ()
71
import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=lowerCAmelCase ) class _lowercase ( lowerCAmelCase ): """simple docstring""" __A = field(default="automatic-speech-recognition", metadata={"include_in_asdict_even_if_is_default": True} ) __A = Features({"audio": Audio()} ) __A = Features({"transcription": Value("string" )} ) __A = "audio" __A = "transcription" def UpperCamelCase_ (self , lowerCamelCase_ ): """simple docstring""" if self.audio_column not in features: raise ValueError(F'''Column {self.audio_column} is not present in features.''' ) if not isinstance(features[self.audio_column] , lowerCamelCase_ ): raise ValueError(F'''Column {self.audio_column} is not an Audio type.''' ) a = copy.deepcopy(self ) a = self.input_schema.copy() a = features[self.audio_column] a = input_schema return task_template @property def UpperCamelCase_ (self ): """simple docstring""" return {self.audio_column: "audio", self.transcription_column: "transcription"}
71
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { """vinvino02/glpn-kitti""": """https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json""", # See all GLPN models at https://huggingface.co/models?filter=glpn } class _lowerCamelCase ( a_ ): _lowerCamelCase :Any = "glpn" def __init__( self : int , UpperCamelCase : Optional[int]=3 , UpperCamelCase : Optional[int]=4 , UpperCamelCase : Union[str, Any]=[2, 2, 2, 2] , UpperCamelCase : List[str]=[8, 4, 2, 1] , UpperCamelCase : str=[32, 64, 1_60, 2_56] , UpperCamelCase : Optional[int]=[7, 3, 3, 3] , UpperCamelCase : Dict=[4, 2, 2, 2] , UpperCamelCase : Tuple=[1, 2, 5, 8] , UpperCamelCase : Optional[Any]=[4, 4, 4, 4] , UpperCamelCase : Dict="gelu" , UpperCamelCase : Dict=0.0 , UpperCamelCase : Optional[int]=0.0 , UpperCamelCase : Union[str, Any]=0.02 , UpperCamelCase : List[str]=0.1 , UpperCamelCase : List[str]=1E-6 , UpperCamelCase : str=64 , UpperCamelCase : Optional[int]=10 , UpperCamelCase : int=-1 , **UpperCamelCase : Any , ) -> Optional[Any]: """simple docstring""" super().__init__(**UpperCamelCase ) lowerCAmelCase__ : int = num_channels lowerCAmelCase__ : Optional[Any] = num_encoder_blocks lowerCAmelCase__ : List[str] = depths lowerCAmelCase__ : Optional[int] = sr_ratios lowerCAmelCase__ : Optional[int] = hidden_sizes lowerCAmelCase__ : Tuple = patch_sizes lowerCAmelCase__ : Optional[int] = strides lowerCAmelCase__ : str = mlp_ratios lowerCAmelCase__ : int = num_attention_heads lowerCAmelCase__ : str = hidden_act lowerCAmelCase__ : Optional[Any] = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : Any = initializer_range lowerCAmelCase__ : int = drop_path_rate lowerCAmelCase__ : Optional[int] = layer_norm_eps lowerCAmelCase__ : int = decoder_hidden_size lowerCAmelCase__ : Optional[int] = max_depth lowerCAmelCase__ : List[Any] = head_in_index
242
"""simple docstring""" import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig _A = logging.get_logger(__name__) class _lowerCamelCase : def __init__( self : Optional[int] , UpperCamelCase : List[Any] , UpperCamelCase : int ) -> str: """simple docstring""" lowerCAmelCase__ : List[Any] = question_encoder lowerCAmelCase__ : Optional[int] = generator lowerCAmelCase__ : Optional[int] = self.question_encoder def _lowerCAmelCase ( self : Dict , UpperCamelCase : Optional[Any] ) -> str: """simple docstring""" if os.path.isfile(UpperCamelCase ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(UpperCamelCase , exist_ok=UpperCamelCase ) lowerCAmelCase__ : Dict = os.path.join(UpperCamelCase , """question_encoder_tokenizer""" ) lowerCAmelCase__ : List[Any] = os.path.join(UpperCamelCase , """generator_tokenizer""" ) self.question_encoder.save_pretrained(UpperCamelCase ) self.generator.save_pretrained(UpperCamelCase ) @classmethod def _lowerCAmelCase ( cls : Union[str, Any] , UpperCamelCase : List[str] , **UpperCamelCase : List[str] ) -> Dict: """simple docstring""" # dynamically import AutoTokenizer from ..auto.tokenization_auto import AutoTokenizer lowerCAmelCase__ : Dict = kwargs.pop("""config""" , UpperCamelCase ) if config is None: lowerCAmelCase__ : int = RagConfig.from_pretrained(UpperCamelCase ) lowerCAmelCase__ : List[str] = AutoTokenizer.from_pretrained( UpperCamelCase , config=config.question_encoder , subfolder="""question_encoder_tokenizer""" ) lowerCAmelCase__ : List[str] = AutoTokenizer.from_pretrained( UpperCamelCase , config=config.generator , subfolder="""generator_tokenizer""" ) return cls(question_encoder=UpperCamelCase , generator=UpperCamelCase ) def __call__( self : Dict , *UpperCamelCase : List[Any] , **UpperCamelCase : Union[str, Any] ) -> int: """simple docstring""" return self.current_tokenizer(*UpperCamelCase , **UpperCamelCase ) def _lowerCAmelCase ( self : Dict , *UpperCamelCase : Tuple , **UpperCamelCase : Optional[int] ) -> Dict: """simple docstring""" return self.generator.batch_decode(*UpperCamelCase , **UpperCamelCase ) def _lowerCAmelCase ( self : List[Any] , *UpperCamelCase : Optional[Any] , **UpperCamelCase : List[Any] ) -> str: """simple docstring""" return self.generator.decode(*UpperCamelCase , **UpperCamelCase ) def _lowerCAmelCase ( self : Any ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase__ : Optional[Any] = self.question_encoder def _lowerCAmelCase ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Optional[int] = self.generator def _lowerCAmelCase ( self : List[str] , UpperCamelCase : List[str] , UpperCamelCase : Optional[List[str]] = None , UpperCamelCase : Optional[int] = None , UpperCamelCase : Optional[int] = None , UpperCamelCase : str = "longest" , UpperCamelCase : str = None , UpperCamelCase : bool = True , **UpperCamelCase : Union[str, Any] , ) -> BatchEncoding: """simple docstring""" warnings.warn( """`prepare_seq2seq_batch` is deprecated and will be removed in version 5 of 🤗 Transformers. Use the """ """regular `__call__` method to prepare your inputs and the tokenizer under the `with_target_tokenizer` """ """context manager to prepare your targets. See the documentation of your specific tokenizer for more """ """details""" , UpperCamelCase , ) if max_length is None: lowerCAmelCase__ : Any = self.current_tokenizer.model_max_length lowerCAmelCase__ : Tuple = self( UpperCamelCase , add_special_tokens=UpperCamelCase , return_tensors=UpperCamelCase , max_length=UpperCamelCase , padding=UpperCamelCase , truncation=UpperCamelCase , **UpperCamelCase , ) if tgt_texts is None: return model_inputs # Process tgt_texts if max_target_length is None: lowerCAmelCase__ : Tuple = self.current_tokenizer.model_max_length lowerCAmelCase__ : Tuple = self( text_target=UpperCamelCase , add_special_tokens=UpperCamelCase , return_tensors=UpperCamelCase , padding=UpperCamelCase , max_length=UpperCamelCase , truncation=UpperCamelCase , **UpperCamelCase , ) lowerCAmelCase__ : Any = labels["""input_ids"""] return model_inputs
242
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __snake_case : List[Any] = { """configuration_data2vec_audio""": ["""DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecAudioConfig"""], """configuration_data2vec_text""": [ """DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecTextConfig""", """Data2VecTextOnnxConfig""", ], """configuration_data2vec_vision""": [ """DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Data2VecVisionConfig""", """Data2VecVisionOnnxConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case : Union[str, Any] = [ """DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecAudioForAudioFrameClassification""", """Data2VecAudioForCTC""", """Data2VecAudioForSequenceClassification""", """Data2VecAudioForXVector""", """Data2VecAudioModel""", """Data2VecAudioPreTrainedModel""", ] __snake_case : Optional[int] = [ """DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecTextForCausalLM""", """Data2VecTextForMaskedLM""", """Data2VecTextForMultipleChoice""", """Data2VecTextForQuestionAnswering""", """Data2VecTextForSequenceClassification""", """Data2VecTextForTokenClassification""", """Data2VecTextModel""", """Data2VecTextPreTrainedModel""", ] __snake_case : Optional[int] = [ """DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST""", """Data2VecVisionForImageClassification""", """Data2VecVisionForMaskedImageModeling""", """Data2VecVisionForSemanticSegmentation""", """Data2VecVisionModel""", """Data2VecVisionPreTrainedModel""", ] if is_tf_available(): __snake_case : List[Any] = [ """TFData2VecVisionForImageClassification""", """TFData2VecVisionForSemanticSegmentation""", """TFData2VecVisionModel""", """TFData2VecVisionPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys __snake_case : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
362
import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __snake_case : Any = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( __lowercase): _SCREAMING_SNAKE_CASE : Tuple = ['''input_features''', '''is_longer'''] def __init__( self , _UpperCamelCase=64 , _UpperCamelCase=4_80_00 , _UpperCamelCase=4_80 , _UpperCamelCase=10 , _UpperCamelCase=10_24 , _UpperCamelCase=0.0 , _UpperCamelCase=False , _UpperCamelCase = 0 , _UpperCamelCase = 1_40_00 , _UpperCamelCase = None , _UpperCamelCase = "fusion" , _UpperCamelCase = "repeatpad" , **_UpperCamelCase , ): """simple docstring""" super().__init__( feature_size=_UpperCamelCase , sampling_rate=_UpperCamelCase , padding_value=_UpperCamelCase , return_attention_mask=_UpperCamelCase , **_UpperCamelCase , ) lowerCAmelCase__ = top_db lowerCAmelCase__ = truncation lowerCAmelCase__ = padding lowerCAmelCase__ = fft_window_size lowerCAmelCase__ = (fft_window_size >> 1) + 1 lowerCAmelCase__ = hop_length lowerCAmelCase__ = max_length_s lowerCAmelCase__ = max_length_s * sampling_rate lowerCAmelCase__ = sampling_rate lowerCAmelCase__ = frequency_min lowerCAmelCase__ = frequency_max lowerCAmelCase__ = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=_UpperCamelCase , min_frequency=_UpperCamelCase , max_frequency=_UpperCamelCase , sampling_rate=_UpperCamelCase , norm=_UpperCamelCase , mel_scale='htk' , ) lowerCAmelCase__ = mel_filter_bank( num_frequency_bins=self.nb_frequency_bins , num_mel_filters=_UpperCamelCase , min_frequency=_UpperCamelCase , max_frequency=_UpperCamelCase , sampling_rate=_UpperCamelCase , norm='slaney' , mel_scale='slaney' , ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase__ = copy.deepcopy(self.__dict__ ) lowerCAmelCase__ = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] if "mel_filters_slaney" in output: del output["mel_filters_slaney"] return output def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None ): """simple docstring""" lowerCAmelCase__ = spectrogram( _UpperCamelCase , window_function(self.fft_window_size , 'hann' ) , frame_length=self.fft_window_size , hop_length=self.hop_length , power=2.0 , mel_filters=_UpperCamelCase , log_mel='dB' , ) return log_mel_spectrogram.T def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = np.array_split(list(range(0 , total_frames - chunk_frames + 1 ) ) , 3 ) if len(ranges[1] ) == 0: # if the audio is too short, we just use the first chunk lowerCAmelCase__ = [0] if len(ranges[2] ) == 0: # if the audio is too short, we just use the first chunk lowerCAmelCase__ = [0] # randomly choose index for each part lowerCAmelCase__ = np.random.choice(ranges[0] ) lowerCAmelCase__ = np.random.choice(ranges[1] ) lowerCAmelCase__ = np.random.choice(ranges[2] ) lowerCAmelCase__ = mel[idx_front : idx_front + chunk_frames, :] lowerCAmelCase__ = mel[idx_middle : idx_middle + chunk_frames, :] lowerCAmelCase__ = mel[idx_back : idx_back + chunk_frames, :] lowerCAmelCase__ = torch.tensor(mel[None, None, :] ) lowerCAmelCase__ = torch.nn.functional.interpolate( _UpperCamelCase , size=[chunk_frames, 64] , mode='bilinear' , align_corners=_UpperCamelCase ) lowerCAmelCase__ = mel_shrink[0][0].numpy() lowerCAmelCase__ = np.stack([mel_shrink, mel_chunk_front, mel_chunk_middle, mel_chunk_back] , axis=0 ) return mel_fusion def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" if waveform.shape[0] > max_length: if truncation == "rand_trunc": lowerCAmelCase__ = True # random crop to max_length (for compatibility) -> this should be handled by self.pad lowerCAmelCase__ = len(_UpperCamelCase ) - max_length lowerCAmelCase__ = np.random.randint(0 , overflow + 1 ) lowerCAmelCase__ = waveform[idx : idx + max_length] lowerCAmelCase__ = self._np_extract_fbank_features(_UpperCamelCase , self.mel_filters_slaney )[None, :] elif truncation == "fusion": lowerCAmelCase__ = self._np_extract_fbank_features(_UpperCamelCase , self.mel_filters ) lowerCAmelCase__ = max_length // self.hop_length + 1 # the +1 related to how the spectrogram is computed lowerCAmelCase__ = mel.shape[0] if chunk_frames == total_frames: # there is a corner case where the audio length is larger than max_length but smaller than max_length+hop_length. # In this case, we just use the whole audio. lowerCAmelCase__ = np.stack([mel, mel, mel, mel] , axis=0 ) lowerCAmelCase__ = False else: lowerCAmelCase__ = self._random_mel_fusion(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) lowerCAmelCase__ = True else: raise NotImplementedError(F"data_truncating {truncation} not implemented" ) else: lowerCAmelCase__ = False # only use repeat as a new possible value for padding. you repeat the audio before applying the usual max_length padding if waveform.shape[0] < max_length: if padding == "repeat": lowerCAmelCase__ = int(max_length / len(_UpperCamelCase ) ) lowerCAmelCase__ = np.stack(np.tile(_UpperCamelCase , n_repeat + 1 ) )[:max_length] if padding == "repeatpad": lowerCAmelCase__ = int(max_length / len(_UpperCamelCase ) ) lowerCAmelCase__ = np.stack(np.tile(_UpperCamelCase , _UpperCamelCase ) ) lowerCAmelCase__ = np.pad(_UpperCamelCase , (0, max_length - waveform.shape[0]) , mode='constant' , constant_values=0 ) if truncation == "fusion": lowerCAmelCase__ = self._np_extract_fbank_features(_UpperCamelCase , self.mel_filters ) lowerCAmelCase__ = np.stack([input_mel, input_mel, input_mel, input_mel] , axis=0 ) else: lowerCAmelCase__ = self._np_extract_fbank_features(_UpperCamelCase , self.mel_filters_slaney )[None, :] return input_mel, longer def __call__( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = None , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = truncation if truncation is not None else self.truncation lowerCAmelCase__ = padding if padding else self.padding if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a" F" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input" F" was sampled with {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) lowerCAmelCase__ = isinstance(_UpperCamelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F"Only mono-channel audio is supported for input to {self}" ) lowerCAmelCase__ = is_batched_numpy or ( isinstance(_UpperCamelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase__ = [np.asarray(_UpperCamelCase , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(_UpperCamelCase , np.ndarray ): lowerCAmelCase__ = np.asarray(_UpperCamelCase , dtype=np.floataa ) elif isinstance(_UpperCamelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase__ = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase__ = [np.asarray(_UpperCamelCase )] # convert to mel spectrogram, truncate and pad if needed. lowerCAmelCase__ = [ self._get_input_mel(_UpperCamelCase , max_length if max_length else self.nb_max_samples , _UpperCamelCase , _UpperCamelCase ) for waveform in raw_speech ] lowerCAmelCase__ = [] lowerCAmelCase__ = [] for mel, longer in padded_inputs: input_mel.append(_UpperCamelCase ) is_longer.append(_UpperCamelCase ) if truncation == "fusion" and sum(_UpperCamelCase ) == 0: # if no audio is longer than 10s, then randomly select one audio to be longer lowerCAmelCase__ = np.random.randint(0 , len(_UpperCamelCase ) ) lowerCAmelCase__ = True if isinstance(input_mel[0] , _UpperCamelCase ): lowerCAmelCase__ = [np.asarray(_UpperCamelCase , dtype=np.floataa ) for feature in input_mel] # is_longer is a list of bool lowerCAmelCase__ = [[longer] for longer in is_longer] lowerCAmelCase__ = {'input_features': input_mel, 'is_longer': is_longer} lowerCAmelCase__ = BatchFeature(_UpperCamelCase ) if return_tensors is not None: lowerCAmelCase__ = input_features.convert_to_tensors(_UpperCamelCase ) return input_features
122
0
"""simple docstring""" def _lowercase ( __snake_case ) -> str: return " ".join( "".join(word[::-1] ) if len(__snake_case ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('Hey wollef sroirraw'))
269
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING a__ : List[Any] = logging.get_logger(__name__) a__ : Union[str, Any] = { "Salesforce/instruct-blip-flan-t5": "https://huggingface.co/Salesforce/instruct-blip-flan-t5/resolve/main/config.json", } class UpperCamelCase__ ( SCREAMING_SNAKE_CASE): UpperCAmelCase__ : List[str] = 'instructblip_vision_model' def __init__( self :List[str] , _A :str=1_408 , _A :List[str]=6_144 , _A :List[Any]=39 , _A :Optional[Any]=16 , _A :Tuple=224 , _A :Tuple=14 , _A :Tuple="gelu" , _A :Optional[Any]=1E-6 , _A :List[Any]=0.0 , _A :Dict=1E-10 , _A :List[str]=True , **_A :Dict , ) -> Dict: '''simple docstring''' super().__init__(**_A ) __A = hidden_size __A = intermediate_size __A = num_hidden_layers __A = num_attention_heads __A = patch_size __A = image_size __A = initializer_range __A = attention_dropout __A = layer_norm_eps __A = hidden_act __A = qkv_bias @classmethod def lowercase_ ( cls :Any , _A :Union[str, os.PathLike] , **_A :Tuple ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(_A ) __A , __A = cls.get_config_dict(_A , **_A ) # get the vision config dict if we are loading from InstructBlipConfig if config_dict.get('model_type' ) == "instructblip": __A = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(_A , **_A ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE): UpperCAmelCase__ : List[str] = 'instructblip_qformer' def __init__( self :Tuple , _A :int=30_522 , _A :List[str]=768 , _A :str=12 , _A :Optional[Any]=12 , _A :Union[str, Any]=3_072 , _A :str="gelu" , _A :Tuple=0.1 , _A :Dict=0.1 , _A :Dict=512 , _A :Union[str, Any]=0.02 , _A :int=1E-12 , _A :str=0 , _A :Union[str, Any]="absolute" , _A :List[str]=2 , _A :Optional[Any]=1_408 , **_A :Any , ) -> Optional[Any]: '''simple docstring''' super().__init__(pad_token_id=_A , **_A ) __A = vocab_size __A = hidden_size __A = num_hidden_layers __A = num_attention_heads __A = hidden_act __A = intermediate_size __A = hidden_dropout_prob __A = attention_probs_dropout_prob __A = max_position_embeddings __A = initializer_range __A = layer_norm_eps __A = position_embedding_type __A = cross_attention_frequency __A = encoder_hidden_size @classmethod def lowercase_ ( cls :int , _A :Union[str, os.PathLike] , **_A :int ) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(_A ) __A , __A = cls.get_config_dict(_A , **_A ) # get the qformer config dict if we are loading from InstructBlipConfig if config_dict.get('model_type' ) == "instructblip": __A = config_dict['qformer_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( F'You are using a model of type {config_dict["model_type"]} to instantiate a model of type ' F'{cls.model_type}. This is not supported for all configurations of models and can yield errors.' ) return cls.from_dict(_A , **_A ) class UpperCamelCase__ ( SCREAMING_SNAKE_CASE): UpperCAmelCase__ : Any = 'instructblip' UpperCAmelCase__ : List[Any] = True def __init__( self :Dict , _A :int=None , _A :Optional[Any]=None , _A :Optional[Any]=None , _A :Optional[Any]=32 , **_A :List[Any] ) -> Tuple: '''simple docstring''' super().__init__(**_A ) if vision_config is None: __A = {} logger.info('vision_config is None. initializing the InstructBlipVisionConfig with default values.' ) if qformer_config is None: __A = {} logger.info('qformer_config is None. Initializing the InstructBlipQFormerConfig with default values.' ) if text_config is None: __A = {} logger.info('text_config is None. Initializing the text config with default values (`OPTConfig`).' ) __A = InstructBlipVisionConfig(**_A ) __A = InstructBlipQFormerConfig(**_A ) __A = text_config['model_type'] if 'model_type' in text_config else 'opt' __A = CONFIG_MAPPING[text_model_type](**_A ) __A = self.text_config.tie_word_embeddings __A = self.text_config.is_encoder_decoder __A = num_query_tokens __A = self.vision_config.hidden_size __A = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES __A = 1.0 __A = 0.02 @classmethod def lowercase_ ( cls :int , _A :InstructBlipVisionConfig , _A :InstructBlipQFormerConfig , _A :PretrainedConfig , **_A :Any , ) -> Any: '''simple docstring''' return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **_A , ) def lowercase_ ( self :int ) -> Tuple: '''simple docstring''' __A = copy.deepcopy(self.__dict__ ) __A = self.vision_config.to_dict() __A = self.qformer_config.to_dict() __A = self.text_config.to_dict() __A = self.__class__.model_type return output
161
0
'''simple docstring''' import functools def _A ( _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =len(_lowerCAmelCase ) __lowercase =len(_lowerCAmelCase ) @functools.cache def min_distance(_lowerCAmelCase , _lowerCAmelCase ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa __lowercase =int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 , _lowerCAmelCase ) , 1 + min_distance(_lowerCAmelCase , indexa + 1 ) , diff + min_distance(indexa + 1 , indexa + 1 ) , ) return min_distance(0 , 0 ) if __name__ == "__main__": import doctest doctest.testmod()
362
'''simple docstring''' import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { """b0""": efficientnet.EfficientNetBa, """b1""": efficientnet.EfficientNetBa, """b2""": efficientnet.EfficientNetBa, """b3""": efficientnet.EfficientNetBa, """b4""": efficientnet.EfficientNetBa, """b5""": efficientnet.EfficientNetBa, """b6""": efficientnet.EfficientNetBa, """b7""": efficientnet.EfficientNetBa, } lowerCamelCase = { """b0""": { """hidden_dim""": 1280, """width_coef""": 1.0, """depth_coef""": 1.0, """image_size""": 224, """dropout_rate""": 0.2, """dw_padding""": [], }, """b1""": { """hidden_dim""": 1280, """width_coef""": 1.0, """depth_coef""": 1.1, """image_size""": 240, """dropout_rate""": 0.2, """dw_padding""": [16], }, """b2""": { """hidden_dim""": 1408, """width_coef""": 1.1, """depth_coef""": 1.2, """image_size""": 260, """dropout_rate""": 0.3, """dw_padding""": [5, 8, 16], }, """b3""": { """hidden_dim""": 1536, """width_coef""": 1.2, """depth_coef""": 1.4, """image_size""": 300, """dropout_rate""": 0.3, """dw_padding""": [5, 18], }, """b4""": { """hidden_dim""": 1792, """width_coef""": 1.4, """depth_coef""": 1.8, """image_size""": 380, """dropout_rate""": 0.4, """dw_padding""": [6], }, """b5""": { """hidden_dim""": 2048, """width_coef""": 1.6, """depth_coef""": 2.2, """image_size""": 456, """dropout_rate""": 0.4, """dw_padding""": [13, 27], }, """b6""": { """hidden_dim""": 2304, """width_coef""": 1.8, """depth_coef""": 2.6, """image_size""": 528, """dropout_rate""": 0.5, """dw_padding""": [31], }, """b7""": { """hidden_dim""": 2560, """width_coef""": 2.0, """depth_coef""": 3.1, """image_size""": 600, """dropout_rate""": 0.5, """dw_padding""": [18], }, } def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =EfficientNetConfig() __lowercase =CONFIG_MAP[model_name]['hidden_dim'] __lowercase =CONFIG_MAP[model_name]['width_coef'] __lowercase =CONFIG_MAP[model_name]['depth_coef'] __lowercase =CONFIG_MAP[model_name]['image_size'] __lowercase =CONFIG_MAP[model_name]['dropout_rate'] __lowercase =CONFIG_MAP[model_name]['dw_padding'] __lowercase ='huggingface/label-files' __lowercase ='imagenet-1k-id2label.json' __lowercase =1_000 __lowercase =json.load(open(hf_hub_download(_lowerCAmelCase , _lowerCAmelCase , repo_type='dataset' ) , 'r' ) ) __lowercase ={int(_lowerCAmelCase ): v for k, v in idalabel.items()} __lowercase =idalabel __lowercase ={v: k for k, v in idalabel.items()} return config def _A ( ): """simple docstring""" __lowercase ='http://images.cocodataset.org/val2017/000000039769.jpg' __lowercase =Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =CONFIG_MAP[model_name]['image_size'] __lowercase =EfficientNetImageProcessor( size={'height': size, 'width': size} , image_mean=[0.4_85, 0.4_56, 0.4_06] , image_std=[0.47_85_39_44, 0.4_73_28_64, 0.47_43_41_63] , do_center_crop=_lowerCAmelCase , ) return preprocessor def _A ( _lowerCAmelCase ): """simple docstring""" __lowercase =[v.split('_' )[0].split('block' )[1] for v in original_param_names if v.startswith('block' )] __lowercase =sorted(set(_lowerCAmelCase ) ) __lowercase =len(_lowerCAmelCase ) __lowercase ={b: str(_lowerCAmelCase ) for b, i in zip(_lowerCAmelCase , range(_lowerCAmelCase ) )} __lowercase =[] rename_keys.append(('stem_conv/kernel:0', 'embeddings.convolution.weight') ) rename_keys.append(('stem_bn/gamma:0', 'embeddings.batchnorm.weight') ) rename_keys.append(('stem_bn/beta:0', 'embeddings.batchnorm.bias') ) rename_keys.append(('stem_bn/moving_mean:0', 'embeddings.batchnorm.running_mean') ) rename_keys.append(('stem_bn/moving_variance:0', 'embeddings.batchnorm.running_var') ) for b in block_names: __lowercase =block_name_mapping[b] rename_keys.append((f"""block{b}_expand_conv/kernel:0""", f"""encoder.blocks.{hf_b}.expansion.expand_conv.weight""") ) rename_keys.append((f"""block{b}_expand_bn/gamma:0""", f"""encoder.blocks.{hf_b}.expansion.expand_bn.weight""") ) rename_keys.append((f"""block{b}_expand_bn/beta:0""", f"""encoder.blocks.{hf_b}.expansion.expand_bn.bias""") ) rename_keys.append( (f"""block{b}_expand_bn/moving_mean:0""", f"""encoder.blocks.{hf_b}.expansion.expand_bn.running_mean""") ) rename_keys.append( (f"""block{b}_expand_bn/moving_variance:0""", f"""encoder.blocks.{hf_b}.expansion.expand_bn.running_var""") ) rename_keys.append( (f"""block{b}_dwconv/depthwise_kernel:0""", f"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight""") ) rename_keys.append((f"""block{b}_bn/gamma:0""", f"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight""") ) rename_keys.append((f"""block{b}_bn/beta:0""", f"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias""") ) rename_keys.append( (f"""block{b}_bn/moving_mean:0""", f"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean""") ) rename_keys.append( (f"""block{b}_bn/moving_variance:0""", f"""encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var""") ) rename_keys.append((f"""block{b}_se_reduce/kernel:0""", f"""encoder.blocks.{hf_b}.squeeze_excite.reduce.weight""") ) rename_keys.append((f"""block{b}_se_reduce/bias:0""", f"""encoder.blocks.{hf_b}.squeeze_excite.reduce.bias""") ) rename_keys.append((f"""block{b}_se_expand/kernel:0""", f"""encoder.blocks.{hf_b}.squeeze_excite.expand.weight""") ) rename_keys.append((f"""block{b}_se_expand/bias:0""", f"""encoder.blocks.{hf_b}.squeeze_excite.expand.bias""") ) rename_keys.append( (f"""block{b}_project_conv/kernel:0""", f"""encoder.blocks.{hf_b}.projection.project_conv.weight""") ) rename_keys.append((f"""block{b}_project_bn/gamma:0""", f"""encoder.blocks.{hf_b}.projection.project_bn.weight""") ) rename_keys.append((f"""block{b}_project_bn/beta:0""", f"""encoder.blocks.{hf_b}.projection.project_bn.bias""") ) rename_keys.append( (f"""block{b}_project_bn/moving_mean:0""", f"""encoder.blocks.{hf_b}.projection.project_bn.running_mean""") ) rename_keys.append( (f"""block{b}_project_bn/moving_variance:0""", f"""encoder.blocks.{hf_b}.projection.project_bn.running_var""") ) rename_keys.append(('top_conv/kernel:0', 'encoder.top_conv.weight') ) rename_keys.append(('top_bn/gamma:0', 'encoder.top_bn.weight') ) rename_keys.append(('top_bn/beta:0', 'encoder.top_bn.bias') ) rename_keys.append(('top_bn/moving_mean:0', 'encoder.top_bn.running_mean') ) rename_keys.append(('top_bn/moving_variance:0', 'encoder.top_bn.running_var') ) __lowercase ={} for item in rename_keys: if item[0] in original_param_names: __lowercase ='efficientnet.' + item[1] __lowercase ='classifier.weight' __lowercase ='classifier.bias' return key_mapping def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" for key, value in tf_params.items(): if "normalization" in key: continue __lowercase =key_mapping[key] if "_conv" in key and "kernel" in key: __lowercase =torch.from_numpy(_lowerCAmelCase ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: __lowercase =torch.from_numpy(_lowerCAmelCase ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: __lowercase =torch.from_numpy(np.transpose(_lowerCAmelCase ) ) else: __lowercase =torch.from_numpy(_lowerCAmelCase ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(_lowerCAmelCase ) @torch.no_grad() def _A ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __lowercase =model_classes[model_name]( include_top=_lowerCAmelCase , weights='imagenet' , input_tensor=_lowerCAmelCase , input_shape=_lowerCAmelCase , pooling=_lowerCAmelCase , classes=1_000 , classifier_activation='softmax' , ) __lowercase =original_model.trainable_variables __lowercase =original_model.non_trainable_variables __lowercase ={param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: __lowercase =param.numpy() __lowercase =list(tf_params.keys() ) # Load HuggingFace model __lowercase =get_efficientnet_config(_lowerCAmelCase ) __lowercase =EfficientNetForImageClassification(_lowerCAmelCase ).eval() __lowercase =hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print('Converting parameters...' ) __lowercase =rename_keys(_lowerCAmelCase ) replace_params(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Initialize preprocessor and preprocess input image __lowercase =convert_image_processor(_lowerCAmelCase ) __lowercase =preprocessor(images=prepare_img() , return_tensors='pt' ) # HF model inference hf_model.eval() with torch.no_grad(): __lowercase =hf_model(**_lowerCAmelCase ) __lowercase =outputs.logits.detach().numpy() # Original model inference __lowercase =False __lowercase =CONFIG_MAP[model_name]['image_size'] __lowercase =prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) __lowercase =image.img_to_array(_lowerCAmelCase ) __lowercase =np.expand_dims(_lowerCAmelCase , axis=0 ) __lowercase =original_model.predict(_lowerCAmelCase ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1e-3 ), "The predicted logits are not the same." print('Model outputs match!' ) if save_model: # Create folder to save model if not os.path.isdir(_lowerCAmelCase ): os.mkdir(_lowerCAmelCase ) # Save converted model and image processor hf_model.save_pretrained(_lowerCAmelCase ) preprocessor.save_pretrained(_lowerCAmelCase ) if push_to_hub: # Push model and image processor to hub print(f"""Pushing converted {model_name} to the hub...""" ) __lowercase =f"""efficientnet-{model_name}""" preprocessor.push_to_hub(_lowerCAmelCase ) hf_model.push_to_hub(_lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""b0""", type=str, help="""Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""hf_model""", type=str, help="""Path to the output PyTorch model directory.""", ) parser.add_argument("""--save_model""", action="""store_true""", help="""Save model to local""") parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") lowerCamelCase = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
48
0
'''simple docstring''' from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class lowerCamelCase ( lowercase_ ): '''simple docstring''' def __init__( self : List[Any] , lowerCAmelCase_ : pyspark.sql.DataFrame , lowerCAmelCase_ : Optional[NamedSplit] = None , lowerCAmelCase_ : Optional[Features] = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : str = None , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : str = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : str = "arrow" , **lowerCAmelCase_ : Dict , ) -> Tuple: '''simple docstring''' super().__init__( split=lowerCAmelCase_ , features=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ , streaming=lowerCAmelCase_ , **lowerCAmelCase_ , ) A__ : Optional[Any] =load_from_cache_file A__ : Optional[int] =file_format A__ : Tuple =Spark( df=lowerCAmelCase_ , features=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , working_dir=lowerCAmelCase_ , **lowerCAmelCase_ , ) def lowercase__ ( self : str ) -> List[Any]: '''simple docstring''' if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) A__ : Union[str, Any] =None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=lowerCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
134
'''simple docstring''' import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class lowerCamelCase ( unittest.TestCase , lowercase_ ): '''simple docstring''' def lowercase__ ( self : int ) -> Any: '''simple docstring''' A__ : int =load_tool("""text-to-speech""" ) self.tool.setup() def lowercase__ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' # SpeechT5 isn't deterministic torch.manual_seed(0 ) A__ : List[str] =self.tool("""hey""" ) A__ : Dict =result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485] ) , ) ) def lowercase__ ( self : Any ) -> Tuple: '''simple docstring''' # SpeechT5 isn't deterministic torch.manual_seed(0 ) A__ : Optional[int] =self.tool("""hey""" ) A__ : Tuple =result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0005966668832115829, -0.0003657640190795064, -0.00013439502799883485] ) , ) )
134
1
def __lowerCamelCase ( __a :int = 3 , __a :int = 7 , __a :int = 1_0_0_0_0_0_0 ) -> Optional[Any]: """simple docstring""" A__ = 0 A__ = 1 for current_denominator in range(1 , limit + 1 ): A__ = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: A__ = current_numerator A__ = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1_0_0_0_0_0_0))
369
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import EsmForMaskedLM, EsmForSequenceClassification, EsmForTokenClassification, EsmModel from transformers.models.esm.modeling_esm import ( ESM_PRETRAINED_MODEL_ARCHIVE_LIST, EsmEmbeddings, create_position_ids_from_input_ids, ) class A : '''simple docstring''' def __init__( self : Union[str, Any] , __lowerCAmelCase : int , __lowerCAmelCase : Tuple=13 , __lowerCAmelCase : Optional[Any]=7 , __lowerCAmelCase : List[str]=False , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : Tuple=False , __lowerCAmelCase : Any=True , __lowerCAmelCase : Union[str, Any]=33 , __lowerCAmelCase : List[str]=32 , __lowerCAmelCase : Optional[Any]=5 , __lowerCAmelCase : Dict=4 , __lowerCAmelCase : List[Any]=37 , __lowerCAmelCase : str="gelu" , __lowerCAmelCase : Any=0.1 , __lowerCAmelCase : str=0.1 , __lowerCAmelCase : List[Any]=5_12 , __lowerCAmelCase : Dict=16 , __lowerCAmelCase : Any=2 , __lowerCAmelCase : List[str]=0.0_2 , __lowerCAmelCase : Dict=3 , __lowerCAmelCase : Optional[int]=4 , __lowerCAmelCase : Tuple=None , ) -> int: """simple docstring""" A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_token_type_ids A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = type_sequence_label_size A__ = initializer_range A__ = num_labels A__ = num_choices A__ = scope def a_ ( self : List[Any] ) -> Tuple: """simple docstring""" A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) A__ = None A__ = None A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ = ids_tensor([self.batch_size] , self.num_choices ) A__ = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def a_ ( self : Optional[int] ) -> str: """simple docstring""" return EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , pad_token_id=1 , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) def a_ ( self : str , __lowerCAmelCase : List[Any] , __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : Any , __lowerCAmelCase : Optional[int] ) -> str: """simple docstring""" A__ = EsmModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() A__ = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase ) A__ = model(__lowerCAmelCase ) A__ = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a_ ( self : List[Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Any ) -> str: """simple docstring""" A__ = EsmForMaskedLM(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() A__ = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a_ ( self : Optional[int] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : int , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : str , __lowerCAmelCase : str , __lowerCAmelCase : List[str] ) -> Any: """simple docstring""" A__ = self.num_labels A__ = EsmForTokenClassification(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() A__ = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a_ ( self : Any ) -> Dict: """simple docstring""" A__ = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) = config_and_inputs A__ = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class A (SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' __lowerCamelCase : List[str] = False __lowerCamelCase : Union[str, Any] = ( ( EsmForMaskedLM, EsmModel, EsmForSequenceClassification, EsmForTokenClassification, ) if is_torch_available() else () ) __lowerCamelCase : List[Any] = () __lowerCamelCase : Optional[int] = ( { '''feature-extraction''': EsmModel, '''fill-mask''': EsmForMaskedLM, '''text-classification''': EsmForSequenceClassification, '''token-classification''': EsmForTokenClassification, '''zero-shot''': EsmForSequenceClassification, } if is_torch_available() else {} ) __lowerCamelCase : Any = True def a_ ( self : Tuple ) -> Optional[int]: """simple docstring""" A__ = EsmModelTester(self ) A__ = ConfigTester(self , config_class=__lowerCAmelCase , hidden_size=37 ) def a_ ( self : Any ) -> str: """simple docstring""" self.config_tester.run_common_tests() def a_ ( self : List[str] ) -> Optional[int]: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) def a_ ( self : Optional[int] ) -> str: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ = type self.model_tester.create_and_check_model(*__lowerCAmelCase ) def a_ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__lowerCAmelCase ) def a_ ( self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__lowerCAmelCase ) @slow def a_ ( self : Optional[int] ) -> int: """simple docstring""" for model_name in ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = EsmModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def a_ ( self : List[str] ) -> Union[str, Any]: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs()[0] A__ = EsmEmbeddings(config=__lowerCAmelCase ) A__ = torch.as_tensor([[12, 31, 13, model.padding_idx]] ) A__ = torch.as_tensor( [ [ 0 + model.padding_idx + 1, 1 + model.padding_idx + 1, 2 + model.padding_idx + 1, model.padding_idx, ] ] ) A__ = create_position_ids_from_input_ids(__lowerCAmelCase , model.padding_idx ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(__lowerCAmelCase , __lowerCAmelCase ) ) ) def a_ ( self : List[Any] ) -> str: """simple docstring""" A__ = self.model_tester.prepare_config_and_inputs()[0] A__ = EsmEmbeddings(config=__lowerCAmelCase ) A__ = torch.empty(2 , 4 , 30 ) A__ = [ 0 + embeddings.padding_idx + 1, 1 + embeddings.padding_idx + 1, 2 + embeddings.padding_idx + 1, 3 + embeddings.padding_idx + 1, ] A__ = torch.as_tensor([expected_single_positions, expected_single_positions] ) A__ = embeddings.create_position_ids_from_inputs_embeds(__lowerCAmelCase ) self.assertEqual(position_ids.shape , expected_positions.shape ) self.assertTrue(torch.all(torch.eq(__lowerCAmelCase , __lowerCAmelCase ) ) ) @unittest.skip("""Esm does not support embedding resizing""" ) def a_ ( self : Dict ) -> Tuple: """simple docstring""" pass @unittest.skip("""Esm does not support embedding resizing""" ) def a_ ( self : List[str] ) -> Optional[int]: """simple docstring""" pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def a_ ( self : List[Any] ) -> Dict: """simple docstring""" pass @require_torch class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' @slow def a_ ( self : int ) -> Optional[int]: """simple docstring""" with torch.no_grad(): A__ = EsmForMaskedLM.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() A__ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) A__ = model(__lowerCAmelCase )[0] A__ = 33 A__ = torch.Size((1, 6, vocab_size) ) self.assertEqual(output.shape , __lowerCAmelCase ) A__ = torch.tensor( [[[8.9_2_1_5, -1_0.5_8_9_8, -6.4_6_7_1], [-6.3_9_6_7, -1_3.9_1_1_4, -1.1_2_1_2], [-7.7_8_1_2, -1_3.9_5_1_6, -3.7_4_0_6]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowerCAmelCase , atol=1e-4 ) ) @slow def a_ ( self : List[str] ) -> Tuple: """simple docstring""" with torch.no_grad(): A__ = EsmModel.from_pretrained("""facebook/esm2_t6_8M_UR50D""" ) model.eval() A__ = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) A__ = model(__lowerCAmelCase )[0] # compare the actual values for a slice. A__ = torch.tensor( [[[0.1_4_4_4, 0.5_4_1_3, 0.3_2_4_8], [0.3_0_3_4, 0.0_0_5_3, 0.3_1_0_8], [0.3_2_2_8, -0.2_4_9_9, 0.3_4_1_5]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowerCAmelCase , atol=1e-4 ) )
276
0
from __future__ import annotations def __magic_name__ ( __lowerCAmelCase : int , __lowerCAmelCase : int ) -> list[list[int]]: __lowerCamelCase = [] create_all_state(1 , __lowerCAmelCase , __lowerCAmelCase , [] , __lowerCAmelCase ) return result def __magic_name__ ( __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : int , __lowerCAmelCase : list[int] , __lowerCAmelCase : list[list[int]] , ) -> None: if level == 0: total_list.append(current_list[:] ) return for i in range(__lowerCAmelCase , total_number - level + 2 ): current_list.append(__lowerCAmelCase ) create_all_state(i + 1 , __lowerCAmelCase , level - 1 , __lowerCAmelCase , __lowerCAmelCase ) current_list.pop() def __magic_name__ ( __lowerCAmelCase : list[list[int]] ) -> None: for i in total_list: print(*__lowerCAmelCase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Any = 4 SCREAMING_SNAKE_CASE__ : Any = 2 SCREAMING_SNAKE_CASE__ : Optional[Any] = generate_all_combinations(n, k) print_all_state(total_list)
270
from __future__ import annotations from fractions import Fraction def __magic_name__ ( __lowerCAmelCase : int , __lowerCAmelCase : int ) -> bool: return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def __magic_name__ ( __lowerCAmelCase : int ) -> list[str]: __lowerCamelCase = [] __lowerCamelCase = 11 __lowerCamelCase = int('''1''' + '''0''' * digit_len ) for num in range(__lowerCAmelCase , __lowerCAmelCase ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(__lowerCAmelCase , __lowerCAmelCase ): solutions.append(f'''{num}/{den}''' ) den += 1 num += 1 __lowerCamelCase = 10 return solutions def __magic_name__ ( __lowerCAmelCase : int = 2 ) -> int: __lowerCamelCase = 1.0 for fraction in fraction_list(__lowerCAmelCase ): __lowerCamelCase = Fraction(__lowerCAmelCase ) result *= frac.denominator / frac.numerator return int(__lowerCAmelCase ) if __name__ == "__main__": print(solution())
270
1
import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def a__ ( A__, A__ ): SCREAMING_SNAKE_CASE_ : List[Any] = old_name if "patch_embed" in old_name: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Dict = old_name.split('.' ) if layer == "0": SCREAMING_SNAKE_CASE_ : int = old_name.replace('0', 'convolution1' ) elif layer == "1": SCREAMING_SNAKE_CASE_ : Optional[Any] = old_name.replace('1', 'batchnorm_before' ) elif layer == "3": SCREAMING_SNAKE_CASE_ : Dict = old_name.replace('3', 'convolution2' ) else: SCREAMING_SNAKE_CASE_ : Dict = old_name.replace('4', 'batchnorm_after' ) if "network" in old_name and re.search(r'\d\.\d', a__ ): SCREAMING_SNAKE_CASE_ : int = r'\b\d{2}\b' if bool(re.search(a__, a__ ) ): SCREAMING_SNAKE_CASE_ : List[Any] = re.search(r'\d\.\d\d.', a__ ).group() else: SCREAMING_SNAKE_CASE_ : Any = re.search(r'\d\.\d.', a__ ).group() if int(match[0] ) < 6: SCREAMING_SNAKE_CASE_ : Optional[Any] = old_name.replace(a__, '' ) SCREAMING_SNAKE_CASE_ : List[str] = trimmed_name.replace('network', match[0] + '.meta4D_layers.blocks.' + match[2:-1] ) SCREAMING_SNAKE_CASE_ : Tuple = 'intermediate_stages.' + trimmed_name else: SCREAMING_SNAKE_CASE_ : Dict = old_name.replace(a__, '' ) if int(match[2] ) < num_meta4D_last_stage: SCREAMING_SNAKE_CASE_ : Optional[Any] = trimmed_name.replace('network', 'meta4D_layers.blocks.' + match[2] ) else: SCREAMING_SNAKE_CASE_ : int = str(int(match[2] ) - num_meta4D_last_stage ) SCREAMING_SNAKE_CASE_ : Any = trimmed_name.replace('network', 'meta3D_layers.blocks.' + layer_index ) if "norm1" in old_name: SCREAMING_SNAKE_CASE_ : List[Any] = trimmed_name.replace('norm1', 'layernorm1' ) elif "norm2" in old_name: SCREAMING_SNAKE_CASE_ : Any = trimmed_name.replace('norm2', 'layernorm2' ) elif "fc1" in old_name: SCREAMING_SNAKE_CASE_ : Any = trimmed_name.replace('fc1', 'linear_in' ) elif "fc2" in old_name: SCREAMING_SNAKE_CASE_ : Union[str, Any] = trimmed_name.replace('fc2', 'linear_out' ) SCREAMING_SNAKE_CASE_ : Dict = 'last_stage.' + trimmed_name elif "network" in old_name and re.search(r'.\d.', a__ ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = old_name.replace('network', 'intermediate_stages' ) if "fc" in new_name: SCREAMING_SNAKE_CASE_ : List[Any] = new_name.replace('fc', 'convolution' ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): SCREAMING_SNAKE_CASE_ : List[Any] = new_name.replace('norm1', 'batchnorm_before' ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): SCREAMING_SNAKE_CASE_ : Any = new_name.replace('norm2', 'batchnorm_after' ) if "proj" in new_name: SCREAMING_SNAKE_CASE_ : int = new_name.replace('proj', 'projection' ) if "dist_head" in new_name: SCREAMING_SNAKE_CASE_ : str = new_name.replace('dist_head', 'distillation_classifier' ) elif "head" in new_name: SCREAMING_SNAKE_CASE_ : Any = new_name.replace('head', 'classifier' ) elif "patch_embed" in new_name: SCREAMING_SNAKE_CASE_ : Tuple = 'efficientformer.' + new_name elif new_name == "norm.weight" or new_name == "norm.bias": SCREAMING_SNAKE_CASE_ : Optional[Any] = new_name.replace('norm', 'layernorm' ) SCREAMING_SNAKE_CASE_ : Tuple = 'efficientformer.' + new_name else: SCREAMING_SNAKE_CASE_ : Any = 'efficientformer.encoder.' + new_name return new_name def a__ ( A__, A__ ): for key in checkpoint.copy().keys(): SCREAMING_SNAKE_CASE_ : Any = checkpoint.pop(a__ ) SCREAMING_SNAKE_CASE_ : Tuple = val return checkpoint def a__ ( ): SCREAMING_SNAKE_CASE_ : List[Any] = 'http://images.cocodataset.org/val2017/000000039769.jpg' SCREAMING_SNAKE_CASE_ : Optional[int] = Image.open(requests.get(a__, stream=a__ ).raw ) return image def a__ ( A__, A__, A__, A__ ): SCREAMING_SNAKE_CASE_ : Tuple = torch.load(a__, map_location='cpu' )['model'] SCREAMING_SNAKE_CASE_ : List[str] = EfficientFormerConfig.from_json_file(a__ ) SCREAMING_SNAKE_CASE_ : Optional[int] = EfficientFormerForImageClassificationWithTeacher(a__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = '_'.join(checkpoint_path.split('/' )[-1].split('.' )[0].split('_' )[:-1] ) SCREAMING_SNAKE_CASE_ : List[str] = config.depths[-1] - config.num_metaad_blocks + 1 SCREAMING_SNAKE_CASE_ : str = convert_torch_checkpoint(a__, a__ ) model.load_state_dict(a__ ) model.eval() SCREAMING_SNAKE_CASE_ : List[str] = { 'bilinear': PILImageResampling.BILINEAR, 'bicubic': PILImageResampling.BICUBIC, 'nearest': PILImageResampling.NEAREST, } # prepare image SCREAMING_SNAKE_CASE_ : Dict = prepare_img() SCREAMING_SNAKE_CASE_ : List[Any] = 2_5_6 SCREAMING_SNAKE_CASE_ : int = 2_2_4 SCREAMING_SNAKE_CASE_ : str = EfficientFormerImageProcessor( size={'shortest_edge': image_size}, crop_size={'height': crop_size, 'width': crop_size}, resample=pillow_resamplings['bicubic'], ) SCREAMING_SNAKE_CASE_ : int = processor(images=a__, return_tensors='pt' ).pixel_values # original processing pipeline SCREAMING_SNAKE_CASE_ : Optional[int] = Compose( [ Resize(a__, interpolation=pillow_resamplings['bicubic'] ), CenterCrop(a__ ), ToTensor(), Normalize(a__, a__ ), ] ) SCREAMING_SNAKE_CASE_ : Any = image_transforms(a__ ).unsqueeze(0 ) assert torch.allclose(a__, a__ ) SCREAMING_SNAKE_CASE_ : int = model(a__ ) SCREAMING_SNAKE_CASE_ : List[str] = outputs.logits SCREAMING_SNAKE_CASE_ : List[Any] = (1, 1_0_0_0) if "l1" in model_name: SCREAMING_SNAKE_CASE_ : Optional[Any] = torch.Tensor( [-0.13_12, 0.43_53, -1.04_99, -0.51_24, 0.41_83, -0.67_93, -1.37_77, -0.08_93, -0.73_58, -2.43_28] ) assert torch.allclose(logits[0, :1_0], a__, atol=1E-3 ) assert logits.shape == expected_shape elif "l3" in model_name: SCREAMING_SNAKE_CASE_ : int = torch.Tensor( [-1.31_50, -1.54_56, -1.25_56, -0.84_96, -0.71_27, -0.78_97, -0.97_28, -0.30_52, 0.37_51, -0.31_27] ) assert torch.allclose(logits[0, :1_0], a__, atol=1E-3 ) assert logits.shape == expected_shape elif "l7" in model_name: SCREAMING_SNAKE_CASE_ : Optional[int] = torch.Tensor( [-1.02_83, -1.41_31, -0.56_44, -1.31_15, -0.57_85, -1.20_49, -0.75_28, 0.19_92, -0.38_22, -0.08_78] ) assert logits.shape == expected_shape else: raise ValueError( F'''Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7''' ) # Save Checkpoints Path(a__ ).mkdir(exist_ok=a__ ) model.save_pretrained(a__ ) print(F'''Checkpoint successfuly converted. Model saved at {pytorch_dump_path}''' ) processor.save_pretrained(a__ ) print(F'''Processor successfuly saved at {pytorch_dump_path}''' ) if push_to_hub: print('Pushing model to the hub...' ) model.push_to_hub( repo_id=F'''Bearnardd/{pytorch_dump_path}''', commit_message='Add model', use_temp_dir=a__, ) processor.push_to_hub( repo_id=F'''Bearnardd/{pytorch_dump_path}''', commit_message='Add image processor', use_temp_dir=a__, ) if __name__ == "__main__": lowerCAmelCase__ : int =argparse.ArgumentParser() # Required parameters parser.add_argument( '--pytorch_model_path', default=None, type=str, required=True, help='Path to EfficientFormer pytorch checkpoint.', ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The json file for EfficientFormer model config.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Push model and image processor to the hub') parser.add_argument( '--no-push_to_hub', dest='push_to_hub', action='store_false', help='Do not push model and image processor to the hub', ) parser.set_defaults(push_to_hub=True) lowerCAmelCase__ : List[str] =parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
361
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def a__ ( A__ ): if is_torch_version('<', '2.0.0' ) or not hasattr(A__, '_dynamo' ): return False return isinstance(A__, torch._dynamo.eval_frame.OptimizedModule ) def a__ ( A__, A__ = True ): SCREAMING_SNAKE_CASE_ : Optional[Any] = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) SCREAMING_SNAKE_CASE_ : List[str] = is_compiled_module(A__ ) if is_compiled: SCREAMING_SNAKE_CASE_ : List[Any] = model SCREAMING_SNAKE_CASE_ : Dict = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(A__, A__ ): SCREAMING_SNAKE_CASE_ : int = model.module if not keep_fpaa_wrapper: SCREAMING_SNAKE_CASE_ : str = getattr(A__, 'forward' ) SCREAMING_SNAKE_CASE_ : Any = model.__dict__.pop('_original_forward', A__ ) if original_forward is not None: while hasattr(A__, '__wrapped__' ): SCREAMING_SNAKE_CASE_ : Optional[int] = forward.__wrapped__ if forward == original_forward: break SCREAMING_SNAKE_CASE_ : Any = forward if getattr(A__, '_converted_to_transformer_engine', A__ ): convert_model(A__, to_transformer_engine=A__ ) if is_compiled: SCREAMING_SNAKE_CASE_ : List[str] = model SCREAMING_SNAKE_CASE_ : Dict = compiled_model return model def a__ ( ): PartialState().wait_for_everyone() def a__ ( A__, A__ ): if PartialState().distributed_type == DistributedType.TPU: xm.save(A__, A__ ) elif PartialState().local_process_index == 0: torch.save(A__, A__ ) @contextmanager def a__ ( **A__ ): for key, value in kwargs.items(): SCREAMING_SNAKE_CASE_ : List[Any] = str(A__ ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def a__ ( A__ ): if not hasattr(A__, '__qualname__' ) and not hasattr(A__, '__name__' ): SCREAMING_SNAKE_CASE_ : Optional[int] = getattr(A__, '__class__', A__ ) if hasattr(A__, '__qualname__' ): return obj.__qualname__ if hasattr(A__, '__name__' ): return obj.__name__ return str(A__ ) def a__ ( A__, A__ ): for key, value in source.items(): if isinstance(A__, A__ ): SCREAMING_SNAKE_CASE_ : Dict = destination.setdefault(A__, {} ) merge_dicts(A__, A__ ) else: SCREAMING_SNAKE_CASE_ : Tuple = value return destination def a__ ( A__ = None ): if port is None: SCREAMING_SNAKE_CASE_ : Tuple = 2_9_5_0_0 with socket.socket(socket.AF_INET, socket.SOCK_STREAM ) as s: return s.connect_ex(('localhost', port) ) == 0
162
0
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
92
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE from transformers.models.speechta import SpeechTaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.tokenization_utils import AddedToken from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Tuple = get_tests_dir('fixtures/test_sentencepiece_bpe_char.model') @require_sentencepiece @require_tokenizers class __A (snake_case__ , unittest.TestCase): '''simple docstring''' __lowercase: Tuple = SpeechTaTokenizer __lowercase: int = False __lowercase: List[str] = True def lowerCAmelCase ( self : Any ) ->str: """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing snake_case_ = SpeechTaTokenizer(UpperCAmelCase_ ) snake_case_ = AddedToken("""<mask>""" , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) snake_case_ = mask_token tokenizer.add_special_tokens({"""mask_token""": mask_token} ) tokenizer.add_tokens(["""<ctc_blank>"""] ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCAmelCase ( self : Optional[Any] , UpperCAmelCase_ : Optional[Any] ) ->Dict: """simple docstring""" snake_case_ = """this is a test""" snake_case_ = """this is a test""" return input_text, output_text def lowerCAmelCase ( self : str , UpperCAmelCase_ : int , UpperCAmelCase_ : Any=False , UpperCAmelCase_ : Tuple=20 , UpperCAmelCase_ : Dict=5 ) ->List[Any]: """simple docstring""" snake_case_ , snake_case_ = self.get_input_output_texts(UpperCAmelCase_ ) snake_case_ = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) snake_case_ = tokenizer.decode(UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) return text, ids def lowerCAmelCase ( self : Union[str, Any] ) ->Optional[Any]: """simple docstring""" snake_case_ = """<pad>""" snake_case_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase_ ) , UpperCAmelCase_ ) def lowerCAmelCase ( self : int ) ->str: """simple docstring""" snake_case_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-4] , """œ""" ) self.assertEqual(vocab_keys[-2] , """<mask>""" ) self.assertEqual(vocab_keys[-1] , """<ctc_blank>""" ) self.assertEqual(len(UpperCAmelCase_ ) , 81 ) def lowerCAmelCase ( self : Optional[int] ) ->int: """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 79 ) def lowerCAmelCase ( self : Optional[int] ) ->List[Any]: """simple docstring""" snake_case_ = self.get_tokenizers(do_lower_case=UpperCAmelCase_ ) for tokenizer in tokenizers: with self.subTest(F"""{tokenizer.__class__.__name__}""" ): snake_case_ = tokenizer.vocab_size snake_case_ = len(UpperCAmelCase_ ) self.assertNotEqual(UpperCAmelCase_ , 0 ) # We usually have added tokens from the start in tests because our vocab fixtures are # smaller than the original vocabs - let's not assert this # self.assertEqual(vocab_size, all_size) snake_case_ = ["""aaaaa bbbbbb""", """cccccccccdddddddd"""] snake_case_ = tokenizer.add_tokens(UpperCAmelCase_ ) snake_case_ = tokenizer.vocab_size snake_case_ = len(UpperCAmelCase_ ) self.assertNotEqual(UpperCAmelCase_ , 0 ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) self.assertEqual(UpperCAmelCase_ , all_size + len(UpperCAmelCase_ ) ) snake_case_ = tokenizer.encode("""aaaaa bbbbbb low cccccccccdddddddd l""" , add_special_tokens=UpperCAmelCase_ ) self.assertGreaterEqual(len(UpperCAmelCase_ ) , 4 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) snake_case_ = {"""eos_token""": """>>>>|||<||<<|<<""", """pad_token""": """<<<<<|||>|>>>>|>"""} snake_case_ = tokenizer.add_special_tokens(UpperCAmelCase_ ) snake_case_ = tokenizer.vocab_size snake_case_ = len(UpperCAmelCase_ ) self.assertNotEqual(UpperCAmelCase_ , 0 ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , len(UpperCAmelCase_ ) ) self.assertEqual(UpperCAmelCase_ , all_size_a + len(UpperCAmelCase_ ) ) snake_case_ = tokenizer.encode( """>>>>|||<||<<|<< aaaaabbbbbb low cccccccccdddddddd <<<<<|||>|>>>>|> l""" , add_special_tokens=UpperCAmelCase_ ) self.assertGreaterEqual(len(UpperCAmelCase_ ) , 6 ) self.assertGreater(tokens[0] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[0] , tokens[1] ) self.assertGreater(tokens[-3] , tokenizer.vocab_size - 1 ) self.assertGreater(tokens[-3] , tokens[-4] ) self.assertEqual(tokens[0] , tokenizer.eos_token_id ) self.assertEqual(tokens[-3] , tokenizer.pad_token_id ) def lowerCAmelCase ( self : Optional[Any] ) ->Tuple: """simple docstring""" pass def lowerCAmelCase ( self : List[str] ) ->Optional[Any]: """simple docstring""" pass def lowerCAmelCase ( self : List[str] ) ->List[str]: """simple docstring""" snake_case_ = self.get_tokenizer() snake_case_ = tokenizer.tokenize("""This is a test""" ) # fmt: off self.assertListEqual(UpperCAmelCase_ , [SPIECE_UNDERLINE, """T""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """a""", SPIECE_UNDERLINE, """t""", """e""", """s""", """t"""] ) # fmt: on self.assertListEqual( tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) , [4, 32, 11, 10, 12, 4, 10, 12, 4, 7, 4, 6, 5, 12, 6] , ) snake_case_ = tokenizer.tokenize("""I was born in 92000, and this is falsé.""" ) self.assertListEqual( UpperCAmelCase_ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """92000""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) snake_case_ = tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) # fmt: off self.assertListEqual(UpperCAmelCase_ , [4, 30, 4, 20, 7, 12, 4, 25, 8, 13, 9, 4, 10, 9, 4, 3, 23, 4, 7, 9, 14, 4, 6, 11, 10, 12, 4, 10, 12, 4, 19, 7, 15, 12, 73, 26] ) # fmt: on snake_case_ = tokenizer.convert_ids_to_tokens(UpperCAmelCase_ ) self.assertListEqual( UpperCAmelCase_ , [SPIECE_UNDERLINE, """I""", SPIECE_UNDERLINE, """w""", """a""", """s""", SPIECE_UNDERLINE, """b""", """o""", """r""", """n""", SPIECE_UNDERLINE, """i""", """n""", SPIECE_UNDERLINE, """<unk>""", """,""", SPIECE_UNDERLINE, """a""", """n""", """d""", SPIECE_UNDERLINE, """t""", """h""", """i""", """s""", SPIECE_UNDERLINE, """i""", """s""", SPIECE_UNDERLINE, """f""", """a""", """l""", """s""", """é""", """."""] ) @slow def lowerCAmelCase ( self : str ) ->Dict: """simple docstring""" snake_case_ = [ """Transformers (formerly known as pytorch-transformers and pytorch-pretrained-bert) provides """ """general-purpose architectures (BERT, GPT, RoBERTa, XLM, DistilBert, XLNet...) for Natural """ """Language Understanding (NLU) and Natural Language Generation (NLG) with over thirty-two pretrained """ """models in one hundred plus languages and deep interoperability between Jax, PyTorch and TensorFlow.""", """BERT is designed to pre-train deep bidirectional representations from unlabeled text by jointly """ """conditioning on both left and right context in all layers.""", """The quick brown fox jumps over the lazy dog.""", ] # fmt: off snake_case_ = { """input_ids""": [ [4, 32, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 64, 19, 8, 13, 18, 5, 13, 15, 22, 4, 28, 9, 8, 20, 9, 4, 7, 12, 4, 24, 22, 6, 8, 13, 17, 11, 39, 6, 13, 7, 9, 12, 19, 8, 13, 18, 5, 13, 12, 4, 7, 9, 14, 4, 24, 22, 6, 8, 13, 17, 11, 39, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 39, 25, 5, 13, 6, 63, 4, 24, 13, 8, 27, 10, 14, 5, 12, 4, 21, 5, 9, 5, 13, 7, 15, 39, 24, 16, 13, 24, 8, 12, 5, 4, 7, 13, 17, 11, 10, 6, 5, 17, 6, 16, 13, 5, 12, 4, 64, 40, 47, 54, 32, 23, 4, 53, 49, 32, 23, 4, 54, 8, 40, 47, 54, 32, 7, 23, 4, 69, 52, 43, 23, 4, 51, 10, 12, 6, 10, 15, 40, 5, 13, 6, 23, 4, 69, 52, 48, 5, 6, 26, 26, 26, 63, 4, 19, 8, 13, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 61, 9, 14, 5, 13, 12, 6, 7, 9, 14, 10, 9, 21, 4, 64, 48, 52, 61, 63, 4, 7, 9, 14, 4, 48, 7, 6, 16, 13, 7, 15, 4, 52, 7, 9, 21, 16, 7, 21, 5, 4, 53, 5, 9, 5, 13, 7, 6, 10, 8, 9, 4, 64, 48, 52, 53, 63, 4, 20, 10, 6, 11, 4, 8, 27, 5, 13, 4, 6, 11, 10, 13, 6, 22, 39, 6, 20, 8, 4, 24, 13, 5, 6, 13, 7, 10, 9, 5, 14, 4, 18, 8, 14, 5, 15, 12, 4, 10, 9, 4, 8, 9, 5, 4, 11, 16, 9, 14, 13, 5, 14, 4, 24, 15, 16, 12, 4, 15, 7, 9, 21, 16, 7, 21, 5, 12, 4, 7, 9, 14, 4, 14, 5, 5, 24, 4, 10, 9, 6, 5, 13, 8, 24, 5, 13, 7, 25, 10, 15, 10, 6, 22, 4, 25, 5, 6, 20, 5, 5, 9, 4, 58, 7, 37, 23, 4, 49, 22, 32, 8, 13, 17, 11, 4, 7, 9, 14, 4, 32, 5, 9, 12, 8, 13, 55, 15, 8, 20, 26, 2], [4, 40, 47, 54, 32, 4, 10, 12, 4, 14, 5, 12, 10, 21, 9, 5, 14, 4, 6, 8, 4, 24, 13, 5, 39, 6, 13, 7, 10, 9, 4, 14, 5, 5, 24, 4, 25, 10, 14, 10, 13, 5, 17, 6, 10, 8, 9, 7, 15, 4, 13, 5, 24, 13, 5, 12, 5, 9, 6, 7, 6, 10, 8, 9, 12, 4, 19, 13, 8, 18, 4, 16, 9, 15, 7, 25, 5, 15, 5, 14, 4, 6, 5, 37, 6, 4, 25, 22, 4, 46, 8, 10, 9, 6, 15, 22, 4, 17, 8, 9, 14, 10, 6, 10, 8, 9, 10, 9, 21, 4, 8, 9, 4, 25, 8, 6, 11, 4, 15, 5, 19, 6, 4, 7, 9, 14, 4, 13, 10, 21, 11, 6, 4, 17, 8, 9, 6, 5, 37, 6, 4, 10, 9, 4, 7, 15, 15, 4, 15, 7, 22, 5, 13, 12, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [4, 32, 11, 5, 4, 45, 16, 10, 17, 28, 4, 25, 13, 8, 20, 9, 4, 19, 8, 37, 4, 46, 16, 18, 24, 12, 4, 8, 27, 5, 13, 4, 6, 11, 5, 4, 15, 7, 57, 22, 4, 14, 8, 21, 26, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], ], """attention_mask""": [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], ] } # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase_ , model_name="""microsoft/speecht5_asr""" , revision="""c5ef64c71905caeccde0e4462ef3f9077224c524""" , sequences=UpperCAmelCase_ , )
347
0
'''simple docstring''' import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available lowerCAmelCase__ = logging.getLogger(__name__) @dataclass class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : str SCREAMING_SNAKE_CASE : List[str] SCREAMING_SNAKE_CASE : Optional[List[str]] @dataclass class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : List[int] SCREAMING_SNAKE_CASE : List[int] SCREAMING_SNAKE_CASE : Optional[List[int]] = None SCREAMING_SNAKE_CASE : Optional[List[int]] = None class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = 'train' SCREAMING_SNAKE_CASE : List[str] = 'dev' SCREAMING_SNAKE_CASE : Tuple = 'test' class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : Dict ,lowercase__ : Union[Split, str] ): raise NotImplementedError @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : str ): raise NotImplementedError @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : List[InputExample] ,lowercase__ : List[str] ,lowercase__ : int ,lowercase__ : PreTrainedTokenizer ,lowercase__ : List[Any]=False ,lowercase__ : str="[CLS]" ,lowercase__ : List[Any]=1 ,lowercase__ : Any="[SEP]" ,lowercase__ : Any=False ,lowercase__ : Tuple=False ,lowercase__ : Tuple=0 ,lowercase__ : Optional[int]=0 ,lowercase__ : List[Any]=-1_0_0 ,lowercase__ : str=0 ,lowercase__ : Union[str, Any]=True ,): __lowercase = {label: i for i, label in enumerate(lowercase__ )} __lowercase = [] for ex_index, example in enumerate(lowercase__ ): if ex_index % 1_0_0_0_0 == 0: logger.info('''Writing example %d of %d''' ,lowercase__ ,len(lowercase__ ) ) __lowercase = [] __lowercase = [] for word, label in zip(example.words ,example.labels ): __lowercase = tokenizer.tokenize(lowercase__ ) # bert-base-multilingual-cased sometimes output "nothing ([]) when calling tokenize with just a space. if len(lowercase__ ) > 0: tokens.extend(lowercase__ ) # Use the real label id for the first token of the word, and padding ids for the remaining tokens label_ids.extend([label_map[label]] + [pad_token_label_id] * (len(lowercase__ ) - 1) ) # Account for [CLS] and [SEP] with "- 2" and with "- 3" for RoBERTa. __lowercase = tokenizer.num_special_tokens_to_add() if len(lowercase__ ) > max_seq_length - special_tokens_count: __lowercase = tokens[: (max_seq_length - special_tokens_count)] __lowercase = label_ids[: (max_seq_length - special_tokens_count)] # The convention in BERT is: # (a) For sequence pairs: # tokens: [CLS] is this jack ##son ##ville ? [SEP] no it is not . [SEP] # type_ids: 0 0 0 0 0 0 0 0 1 1 1 1 1 1 # (b) For single sequences: # tokens: [CLS] the dog is hairy . [SEP] # type_ids: 0 0 0 0 0 0 0 # # Where "type_ids" are used to indicate whether this is the first # sequence or the second sequence. The embedding vectors for `type=0` and # `type=1` were learned during pre-training and are added to the wordpiece # embedding vector (and position vector). This is not *strictly* necessary # since the [SEP] token unambiguously separates the sequences, but it makes # it easier for the model to learn the concept of sequences. # # For classification tasks, the first vector (corresponding to [CLS]) is # used as the "sentence vector". Note that this only makes sense because # the entire model is fine-tuned. tokens += [sep_token] label_ids += [pad_token_label_id] if sep_token_extra: # roberta uses an extra separator b/w pairs of sentences tokens += [sep_token] label_ids += [pad_token_label_id] __lowercase = [sequence_a_segment_id] * len(lowercase__ ) if cls_token_at_end: tokens += [cls_token] label_ids += [pad_token_label_id] segment_ids += [cls_token_segment_id] else: __lowercase = [cls_token] + tokens __lowercase = [pad_token_label_id] + label_ids __lowercase = [cls_token_segment_id] + segment_ids __lowercase = tokenizer.convert_tokens_to_ids(lowercase__ ) # The mask has 1 for real tokens and 0 for padding tokens. Only real # tokens are attended to. __lowercase = [1 if mask_padding_with_zero else 0] * len(lowercase__ ) # Zero-pad up to the sequence length. __lowercase = max_seq_length - len(lowercase__ ) if pad_on_left: __lowercase = ([pad_token] * padding_length) + input_ids __lowercase = ([0 if mask_padding_with_zero else 1] * padding_length) + input_mask __lowercase = ([pad_token_segment_id] * padding_length) + segment_ids __lowercase = ([pad_token_label_id] * padding_length) + label_ids else: input_ids += [pad_token] * padding_length input_mask += [0 if mask_padding_with_zero else 1] * padding_length segment_ids += [pad_token_segment_id] * padding_length label_ids += [pad_token_label_id] * padding_length assert len(lowercase__ ) == max_seq_length assert len(lowercase__ ) == max_seq_length assert len(lowercase__ ) == max_seq_length assert len(lowercase__ ) == max_seq_length if ex_index < 5: logger.info('''*** Example ***''' ) logger.info('''guid: %s''' ,example.guid ) logger.info('''tokens: %s''' ,''' '''.join([str(lowercase__ ) for x in tokens] ) ) logger.info('''input_ids: %s''' ,''' '''.join([str(lowercase__ ) for x in input_ids] ) ) logger.info('''input_mask: %s''' ,''' '''.join([str(lowercase__ ) for x in input_mask] ) ) logger.info('''segment_ids: %s''' ,''' '''.join([str(lowercase__ ) for x in segment_ids] ) ) logger.info('''label_ids: %s''' ,''' '''.join([str(lowercase__ ) for x in label_ids] ) ) if "token_type_ids" not in tokenizer.model_input_names: __lowercase = None features.append( InputFeatures( input_ids=lowercase__ ,attention_mask=lowercase__ ,token_type_ids=lowercase__ ,label_ids=lowercase__ ) ) return features if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : List[InputFeatures] SCREAMING_SNAKE_CASE : int = nn.CrossEntropyLoss().ignore_index def __init__( self : Union[str, Any] ,lowercase__ : TokenClassificationTask ,lowercase__ : str ,lowercase__ : PreTrainedTokenizer ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Optional[int] = None ,lowercase__ : Dict=False ,lowercase__ : Split = Split.train ,): # Load data features from cache or dataset file __lowercase = os.path.join( lowercase__ ,'''cached_{}_{}_{}'''.format(mode.value ,tokenizer.__class__.__name__ ,str(lowercase__ ) ) ,) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. __lowercase = cached_features_file + '''.lock''' with FileLock(lowercase__ ): if os.path.exists(lowercase__ ) and not overwrite_cache: logger.info(F"Loading features from cached file {cached_features_file}" ) __lowercase = torch.load(lowercase__ ) else: logger.info(F"Creating features from dataset file at {data_dir}" ) __lowercase = token_classification_task.read_examples_from_file(lowercase__ ,lowercase__ ) # TODO clean up all this to leverage built-in features of tokenizers __lowercase = token_classification_task.convert_examples_to_features( lowercase__ ,lowercase__ ,lowercase__ ,lowercase__ ,cls_token_at_end=bool(model_type in ['''xlnet'''] ) ,cls_token=tokenizer.cls_token ,cls_token_segment_id=2 if model_type in ['''xlnet'''] else 0 ,sep_token=tokenizer.sep_token ,sep_token_extra=lowercase__ ,pad_on_left=bool(tokenizer.padding_side == '''left''' ) ,pad_token=tokenizer.pad_token_id ,pad_token_segment_id=tokenizer.pad_token_type_id ,pad_token_label_id=self.pad_token_label_id ,) logger.info(F"Saving features into cached file {cached_features_file}" ) torch.save(self.features ,lowercase__ ) def __len__( self : Tuple ): return len(self.features ) def __getitem__( self : Tuple ,lowercase__ : Optional[int] ): return self.features[i] if is_tf_available(): import tensorflow as tf class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : List[InputFeatures] SCREAMING_SNAKE_CASE : int = -1_0_0 def __init__( self : Any ,lowercase__ : TokenClassificationTask ,lowercase__ : str ,lowercase__ : PreTrainedTokenizer ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Optional[int] = None ,lowercase__ : Union[str, Any]=False ,lowercase__ : Split = Split.train ,): __lowercase = token_classification_task.read_examples_from_file(lowercase__ ,lowercase__ ) # TODO clean up all this to leverage built-in features of tokenizers __lowercase = token_classification_task.convert_examples_to_features( lowercase__ ,lowercase__ ,lowercase__ ,lowercase__ ,cls_token_at_end=bool(model_type in ['''xlnet'''] ) ,cls_token=tokenizer.cls_token ,cls_token_segment_id=2 if model_type in ['''xlnet'''] else 0 ,sep_token=tokenizer.sep_token ,sep_token_extra=lowercase__ ,pad_on_left=bool(tokenizer.padding_side == '''left''' ) ,pad_token=tokenizer.pad_token_id ,pad_token_segment_id=tokenizer.pad_token_type_id ,pad_token_label_id=self.pad_token_label_id ,) def gen(): for ex in self.features: if ex.token_type_ids is None: yield ( {"input_ids": ex.input_ids, "attention_mask": ex.attention_mask}, ex.label_ids, ) else: yield ( { "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label_ids, ) if "token_type_ids" not in tokenizer.model_input_names: __lowercase = tf.data.Dataset.from_generator( lowercase__ ,({'''input_ids''': tf.intaa, '''attention_mask''': tf.intaa}, tf.intaa) ,( {'''input_ids''': tf.TensorShape([None] ), '''attention_mask''': tf.TensorShape([None] )}, tf.TensorShape([None] ), ) ,) else: __lowercase = tf.data.Dataset.from_generator( lowercase__ ,({'''input_ids''': tf.intaa, '''attention_mask''': tf.intaa, '''token_type_ids''': tf.intaa}, tf.intaa) ,( { '''input_ids''': tf.TensorShape([None] ), '''attention_mask''': tf.TensorShape([None] ), '''token_type_ids''': tf.TensorShape([None] ), }, tf.TensorShape([None] ), ) ,) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.dataset.apply(tf.data.experimental.assert_cardinality(len(self.features ) ) ) return self.dataset def __len__( self : int ): return len(self.features ) def __getitem__( self : Optional[Any] ,lowercase__ : List[str] ): return self.features[i]
362
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline lowerCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Dict ,lowercase__ : Tuple ): super().__init__() self.register_modules(unet=lowercase__ ,scheduler=lowercase__ ) @torch.no_grad() def __call__( self : Any ,lowercase__ : int = 1 ,lowercase__ : int = 1_0_0 ,lowercase__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None ,lowercase__ : Optional[float] = None ,lowercase__ : bool = True ,): if audio_length_in_s is None: __lowercase = self.unet.config.sample_size / self.unet.config.sample_rate __lowercase = audio_length_in_s * self.unet.config.sample_rate __lowercase = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( F"{audio_length_in_s} is too small. Make sure it's bigger or equal to" F" {3 * down_scale_factor / self.unet.config.sample_rate}." ) __lowercase = int(lowercase__ ) if sample_size % down_scale_factor != 0: __lowercase = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( F"{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled" F" by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising" ''' process.''' ) __lowercase = int(lowercase__ ) __lowercase = next(iter(self.unet.parameters() ) ).dtype __lowercase = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(lowercase__ ,lowercase__ ) and len(lowercase__ ) != batch_size: raise ValueError( F"You have passed a list of generators of length {len(lowercase__ )}, but requested an effective batch" F" size of {batch_size}. Make sure the batch size matches the length of the generators." ) __lowercase = randn_tensor(lowercase__ ,generator=lowercase__ ,device=self.device ,dtype=lowercase__ ) # set step values self.scheduler.set_timesteps(lowercase__ ,device=audio.device ) __lowercase = self.scheduler.timesteps.to(lowercase__ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output __lowercase = self.unet(lowercase__ ,lowercase__ ).sample # 2. compute previous image: x_t -> t_t-1 __lowercase = self.scheduler.step(lowercase__ ,lowercase__ ,lowercase__ ).prev_sample __lowercase = audio.clamp(-1 ,1 ).float().cpu().numpy() __lowercase = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=lowercase__ )
52
0
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class __lowerCAmelCase ( unittest.TestCase): def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ : Union[str, Any] =0 def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ : Any =AutoImageProcessor.from_pretrained("openai/clip-vit-base-patch32" ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> str: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: a__ : Optional[Any] =Path(lowerCAmelCase__ ) / "preprocessor_config.json" a__ : Union[str, Any] =Path(lowerCAmelCase__ ) / "config.json" json.dump( {"image_processor_type": "CLIPImageProcessor", "processor_class": "CLIPProcessor"} , open(lowerCAmelCase__ , "w" ) , ) json.dump({"model_type": "clip"} , open(lowerCAmelCase__ , "w" ) ) a__ : Any =AutoImageProcessor.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> int: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: a__ : Any =Path(lowerCAmelCase__ ) / "preprocessor_config.json" a__ : Any =Path(lowerCAmelCase__ ) / "config.json" json.dump( {"feature_extractor_type": "CLIPFeatureExtractor", "processor_class": "CLIPProcessor"} , open(lowerCAmelCase__ , "w" ) , ) json.dump({"model_type": "clip"} , open(lowerCAmelCase__ , "w" ) ) a__ : str =AutoImageProcessor.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: a__ : Optional[int] =CLIPConfig() # Create a dummy config file with image_proceesor_type a__ : Tuple =Path(lowerCAmelCase__ ) / "preprocessor_config.json" a__ : Union[str, Any] =Path(lowerCAmelCase__ ) / "config.json" json.dump( {"image_processor_type": "CLIPImageProcessor", "processor_class": "CLIPProcessor"} , open(lowerCAmelCase__ , "w" ) , ) json.dump({"model_type": "clip"} , open(lowerCAmelCase__ , "w" ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally a__ : List[Any] =AutoImageProcessor.from_pretrained(lowerCAmelCase__ ).to_dict() config_dict.pop("image_processor_type" ) a__ : int =CLIPImageProcessor(**lowerCAmelCase__ ) # save in new folder model_config.save_pretrained(lowerCAmelCase__ ) config.save_pretrained(lowerCAmelCase__ ) a__ : Tuple =AutoImageProcessor.from_pretrained(lowerCAmelCase__ ) # make sure private variable is not incorrectly saved a__ : str =json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> List[Any]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: a__ : List[str] =Path(lowerCAmelCase__ ) / "preprocessor_config.json" json.dump( {"image_processor_type": "CLIPImageProcessor", "processor_class": "CLIPProcessor"} , open(lowerCAmelCase__ , "w" ) , ) a__ : Optional[int] =AutoImageProcessor.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) def _lowercase ( self ) -> int: '''simple docstring''' with self.assertRaisesRegex( lowerCAmelCase__ , "clip-base is not a local folder and is not a valid model identifier" ): a__ : Optional[Any] =AutoImageProcessor.from_pretrained("clip-base" ) def _lowercase ( self ) -> Optional[int]: '''simple docstring''' with self.assertRaisesRegex( lowerCAmelCase__ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): a__ : Tuple =AutoImageProcessor.from_pretrained(lowerCAmelCase__ , revision="aaaaaa" ) def _lowercase ( self ) -> int: '''simple docstring''' with self.assertRaisesRegex( lowerCAmelCase__ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): a__ : List[str] =AutoImageProcessor.from_pretrained("hf-internal-testing/config-no-model" ) def _lowercase ( self ) -> Optional[int]: '''simple docstring''' with self.assertRaises(lowerCAmelCase__ ): a__ : Optional[int] =AutoImageProcessor.from_pretrained("hf-internal-testing/test_dynamic_image_processor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowerCAmelCase__ ): a__ : Any =AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=lowerCAmelCase__ ) a__ : Optional[Any] =AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(lowerCAmelCase__ ) a__ : Any =AutoImageProcessor.from_pretrained(lowerCAmelCase__ , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(reloaded_image_processor.__class__.__name__ , "NewImageProcessor" ) def _lowercase ( self ) -> Dict: '''simple docstring''' try: AutoConfig.register("custom" , lowerCAmelCase__ ) AutoImageProcessor.register(lowerCAmelCase__ , lowerCAmelCase__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCAmelCase__ ): AutoImageProcessor.register(lowerCAmelCase__ , lowerCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdirname: a__ : Any =Path(lowerCAmelCase__ ) / "preprocessor_config.json" a__ : Dict =Path(lowerCAmelCase__ ) / "config.json" json.dump( {"feature_extractor_type": "CLIPFeatureExtractor", "processor_class": "CLIPProcessor"} , open(lowerCAmelCase__ , "w" ) , ) json.dump({"model_type": "clip"} , open(lowerCAmelCase__ , "w" ) ) a__ : List[Any] =CustomImageProcessor.from_pretrained(lowerCAmelCase__ ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(lowerCAmelCase__ ) a__ : Tuple =AutoImageProcessor.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ , lowerCAmelCase__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _lowercase ( self ) -> str: '''simple docstring''' class __lowerCAmelCase ( UpperCamelCase__): _lowercase : Union[str, Any] = True try: AutoConfig.register("custom" , lowerCAmelCase__ ) AutoImageProcessor.register(lowerCAmelCase__ , lowerCAmelCase__ ) # If remote code is not set, the default is to use local a__ : List[str] =AutoImageProcessor.from_pretrained("hf-internal-testing/test_dynamic_image_processor" ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. a__ : str =AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub a__ : List[str] =AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=lowerCAmelCase__ ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) self.assertTrue(not hasattr(lowerCAmelCase__ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
95
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase : Optional[Any] = logging.get_logger(__name__) UpperCAmelCase : Optional[int] = { """google/canine-s""": """https://huggingface.co/google/canine-s/resolve/main/config.json""", # See all CANINE models at https://huggingface.co/models?filter=canine } class __lowerCAmelCase ( UpperCamelCase__): _lowercase : List[Any] = """canine""" def __init__( self , lowerCAmelCase__=7_6_8 , lowerCAmelCase__=1_2 , lowerCAmelCase__=1_2 , lowerCAmelCase__=3_0_7_2 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1_6_3_8_4 , lowerCAmelCase__=1_6 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=0 , lowerCAmelCase__=0XE0_00 , lowerCAmelCase__=0XE0_01 , lowerCAmelCase__=4 , lowerCAmelCase__=4 , lowerCAmelCase__=8 , lowerCAmelCase__=1_6_3_8_4 , lowerCAmelCase__=1_2_8 , **lowerCAmelCase__ , ) -> Dict: '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase__ , bos_token_id=lowerCAmelCase__ , eos_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a__ : Optional[int] =max_position_embeddings a__ : str =hidden_size a__ : Optional[Any] =num_hidden_layers a__ : Tuple =num_attention_heads a__ : Optional[Any] =intermediate_size a__ : Optional[int] =hidden_act a__ : List[Any] =hidden_dropout_prob a__ : Union[str, Any] =attention_probs_dropout_prob a__ : Optional[Any] =initializer_range a__ : Union[str, Any] =type_vocab_size a__ : Optional[int] =layer_norm_eps # Character config: a__ : int =downsampling_rate a__ : Optional[Any] =upsampling_kernel_size a__ : Union[str, Any] =num_hash_functions a__ : Any =num_hash_buckets a__ : int =local_transformer_stride
95
1
import os import unittest from transformers.models.cpmant.tokenization_cpmant import VOCAB_FILES_NAMES, CpmAntTokenizer from transformers.testing_utils import require_jieba, tooslow from ...test_tokenization_common import TokenizerTesterMixin @require_jieba class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , unittest.TestCase ): A_ : Optional[Any] = CpmAntTokenizer A_ : Optional[int] = False def a (self : str ): """simple docstring""" super().setUp() __snake_case = [ '''<d>''', '''</d>''', '''<s>''', '''</s>''', '''</_>''', '''<unk>''', '''<pad>''', '''</n>''', '''我''', '''是''', '''C''', '''P''', '''M''', '''A''', '''n''', '''t''', ] __snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) @tooslow def a (self : Tuple ): """simple docstring""" __snake_case = CpmAntTokenizer.from_pretrained('''openbmb/cpm-ant-10b''' ) __snake_case = '''今天天气真好!''' __snake_case = ['''今天''', '''天气''', '''真''', '''好''', '''!'''] __snake_case = tokenizer.tokenize(a__ ) self.assertListEqual(a__ , a__ ) __snake_case = '''今天天气真好!''' __snake_case = [tokenizer.bos_token] + tokens __snake_case = [6, 9802, 1_4962, 2082, 831, 244] self.assertListEqual(tokenizer.convert_tokens_to_ids(a__ ) , a__ ) __snake_case = tokenizer.decode(a__ ) self.assertEqual(a__ , a__ )
238
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case_ = { 'configuration_git': ['GIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GitConfig', 'GitVisionConfig'], 'processing_git': ['GitProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ 'GIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'GitForCausalLM', 'GitModel', 'GitPreTrainedModel', 'GitVisionModel', ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys snake_case_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
238
1
def lowercase__ ( __snake_case : int ): '''simple docstring''' return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
29
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) a_ = {'configuration_encoder_decoder': ['EncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['EncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['TFEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['FlaxEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys a_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
175
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available a__ : Optional[int] = { '''configuration_bloom''': ['''BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BloomConfig''', '''BloomOnnxConfig'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : List[Any] = ['''BloomTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[int] = [ '''BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BloomForCausalLM''', '''BloomModel''', '''BloomPreTrainedModel''', '''BloomForSequenceClassification''', '''BloomForTokenClassification''', '''BloomForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys a__ : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
19
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin a__ : Optional[Any] = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class a_ ( a__ , unittest.TestCase ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = XLMProphetNetTokenizer __SCREAMING_SNAKE_CASE : List[str] = False __SCREAMING_SNAKE_CASE : Dict = True def __lowerCAmelCase ( self ) ->Dict: super().setUp() # We have a SentencePiece fixture for testing SCREAMING_SNAKE_CASE : Optional[Any] = XLMProphetNetTokenizer(_lowerCamelCase , keep_accents=_lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self ) ->Tuple: SCREAMING_SNAKE_CASE : List[str] = '''[PAD]''' SCREAMING_SNAKE_CASE : Tuple = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_lowerCamelCase ) , _lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_lowerCamelCase ) , _lowerCamelCase ) def __lowerCAmelCase ( self ) ->List[Any]: SCREAMING_SNAKE_CASE : Optional[Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''[PAD]''' ) self.assertEqual(vocab_keys[1] , '''[CLS]''' ) self.assertEqual(vocab_keys[-1] , '''j''' ) self.assertEqual(len(_lowerCamelCase ) , 1012 ) def __lowerCAmelCase ( self ) ->List[str]: self.assertEqual(self.get_tokenizer().vocab_size , 1012 ) def __lowerCAmelCase ( self ) ->Optional[int]: SCREAMING_SNAKE_CASE : Union[str, Any] = XLMProphetNetTokenizer(_lowerCamelCase , keep_accents=_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(_lowerCamelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( _lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.convert_tokens_to_ids(_lowerCamelCase ) self.assertListEqual( _lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) SCREAMING_SNAKE_CASE : str = tokenizer.convert_ids_to_tokens(_lowerCamelCase ) self.assertListEqual( _lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''[UNK]''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''[UNK]''', '''.''', ] , ) @cached_property def __lowerCAmelCase ( self ) ->List[str]: return XLMProphetNetTokenizer.from_pretrained('''microsoft/xprophetnet-large-wiki100-cased''' ) @slow def __lowerCAmelCase ( self ) ->Any: SCREAMING_SNAKE_CASE : Union[str, Any] = '''Hello World!''' SCREAMING_SNAKE_CASE : int = [3_5389, 6672, 49, 2] self.assertListEqual(_lowerCamelCase , self.big_tokenizer.encode(_lowerCamelCase ) ) @slow def __lowerCAmelCase ( self ) ->int: # fmt: off SCREAMING_SNAKE_CASE : str = {'''input_ids''': [[1_1073, 8_2783, 18, 26, 8_2783, 549, 5_1540, 248, 1_7209, 1301, 217, 20, 21_5186, 1325, 147, 1_7209, 1301, 217, 20, 5_6370, 53, 12_2020, 20, 1_6477, 27, 8_7355, 4548, 20, 4728, 7_8392, 17, 15_9969, 18, 26, 2_4491, 629, 15, 538, 2_2704, 5439, 15, 2788, 2_4491, 9885, 15, 4_3534, 605, 15, 814, 1_8403, 3_3200, 29, 15, 4_3534, 2_4458, 1_2410, 111, 2_4966, 8_3669, 9637, 14_4068, 26, 850, 2_2346, 27, 147, 2_4966, 8_3669, 8_3490, 26, 3_9113, 735, 27, 689, 656, 2800, 1339, 4600, 53, 12_2020, 11_5785, 34, 816, 1339, 4_6887, 18, 147, 5_3905, 1951, 4_2238, 4_1170, 1_7732, 834, 436, 15, 2_7523, 9_8733, 217, 147, 5542, 4981, 930, 1_7347, 16, 2], [2_0091, 629, 94, 8_2786, 58, 490, 20, 1528, 84, 5_3905, 344, 8_0592, 11_0128, 1_8822, 5267, 1306, 62, 15_2537, 308, 7997, 401, 12_4427, 549, 3_5442, 225, 109, 1_5055, 2_5748, 147, 7119, 4_3712, 34, 767, 13_5366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 6_3784, 11_9466, 17, 14_7808, 8_8214, 18, 656, 81, 32, 3296, 1_0280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_lowerCamelCase , model_name='''microsoft/xprophetnet-large-wiki100-cased''' , revision='''1acad1643ddd54a44df6a1b797ada8373685d90e''' , )
19
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def A ( a_ ) -> Tuple: __UpperCamelCase : int =botoa.client('iam' ) __UpperCamelCase : Any ={ 'Version': '2012-10-17', 'Statement': [ {'Effect': 'Allow', 'Principal': {'Service': 'sagemaker.amazonaws.com'}, 'Action': 'sts:AssumeRole'} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=a_ ,AssumeRolePolicyDocument=json.dumps(a_ ,indent=2 ) ) __UpperCamelCase : Any ={ 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 'sagemaker:*', 'ecr:GetDownloadUrlForLayer', 'ecr:BatchGetImage', 'ecr:BatchCheckLayerAvailability', 'ecr:GetAuthorizationToken', 'cloudwatch:PutMetricData', 'cloudwatch:GetMetricData', 'cloudwatch:GetMetricStatistics', 'cloudwatch:ListMetrics', 'logs:CreateLogGroup', 'logs:CreateLogStream', 'logs:DescribeLogStreams', 'logs:PutLogEvents', 'logs:GetLogEvents', 's3:CreateBucket', 's3:ListBucket', 's3:GetBucketLocation', 's3:GetObject', 's3:PutObject', ], 'Resource': '*', } ], } # attach policy to role iam_client.put_role_policy( RoleName=a_ ,PolicyName=F'{role_name}_policy_permission' ,PolicyDocument=json.dumps(a_ ,indent=2 ) ,) except iam_client.exceptions.EntityAlreadyExistsException: print(F'role {role_name} already exists. Using existing one' ) def A ( a_ ) -> str: __UpperCamelCase : str =botoa.client('iam' ) return iam_client.get_role(RoleName=a_ )["Role"]["Arn"] def A ( ) -> int: __UpperCamelCase : List[str] =_ask_options( 'How do you want to authorize?' ,['AWS Profile', 'Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) '] ,a_ ,) __UpperCamelCase : List[Any] =None if credentials_configuration == 0: __UpperCamelCase : Dict =_ask_field('Enter your AWS Profile name: [default] ' ,default='default' ) __UpperCamelCase : Optional[int] =aws_profile else: print( 'Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,' '`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`' ) __UpperCamelCase : Dict =_ask_field('AWS Access Key ID: ' ) __UpperCamelCase : Optional[int] =aws_access_key_id __UpperCamelCase : Any =_ask_field('AWS Secret Access Key: ' ) __UpperCamelCase : int =aws_secret_access_key __UpperCamelCase : Optional[Any] =_ask_field('Enter your AWS Region: [us-east-1]' ,default='us-east-1' ) __UpperCamelCase : List[Any] =aws_region __UpperCamelCase : List[str] =_ask_options( 'Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?' ,['Provide IAM Role name', 'Create new IAM role using credentials'] ,a_ ,) if role_management == 0: __UpperCamelCase : List[Any] =_ask_field('Enter your IAM role name: ' ) else: __UpperCamelCase : List[str] ='accelerate_sagemaker_execution_role' print(F'Accelerate will create an iam role "{iam_role_name}" using the provided credentials' ) _create_iam_role_for_sagemaker(a_ ) __UpperCamelCase : Dict =_ask_field( 'Do you want to use custom Docker image? [yes/NO]: ' ,_convert_yes_no_to_bool ,default=a_ ,error_message='Please enter yes or no.' ,) __UpperCamelCase : Optional[Any] =None if is_custom_docker_image: __UpperCamelCase : Union[str, Any] =_ask_field('Enter your Docker image: ' ,lambda a_ : str(a_ ).lower() ) __UpperCamelCase : int =_ask_field( 'Do you want to provide SageMaker input channels with data locations? [yes/NO]: ' ,_convert_yes_no_to_bool ,default=a_ ,error_message='Please enter yes or no.' ,) __UpperCamelCase : Optional[int] =None if is_sagemaker_inputs_enabled: __UpperCamelCase : str =_ask_field( 'Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): ' ,lambda a_ : str(a_ ).lower() ,) __UpperCamelCase : int =_ask_field( 'Do you want to enable SageMaker metrics? [yes/NO]: ' ,_convert_yes_no_to_bool ,default=a_ ,error_message='Please enter yes or no.' ,) __UpperCamelCase : List[str] =None if is_sagemaker_metrics_enabled: __UpperCamelCase : str =_ask_field( 'Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): ' ,lambda a_ : str(a_ ).lower() ,) __UpperCamelCase : List[str] =_ask_options( 'What is the distributed mode?' ,['No distributed training', 'Data parallelism'] ,_convert_sagemaker_distributed_mode ,) __UpperCamelCase : List[str] ={} __UpperCamelCase : Optional[int] =_ask_field( 'Do you wish to optimize your script with torch dynamo?[yes/NO]:' ,_convert_yes_no_to_bool ,default=a_ ,error_message='Please enter yes or no.' ,) if use_dynamo: __UpperCamelCase : List[str] ='dynamo_' __UpperCamelCase : int =_ask_options( 'Which dynamo backend would you like to use?' ,[x.lower() for x in DYNAMO_BACKENDS] ,_convert_dynamo_backend ,default=2 ,) __UpperCamelCase : List[Any] =_ask_field( 'Do you want to customize the defaults sent to torch.compile? [yes/NO]: ' ,_convert_yes_no_to_bool ,default=a_ ,error_message='Please enter yes or no.' ,) if use_custom_options: __UpperCamelCase : int =_ask_options( 'Which mode do you want to use?' ,a_ ,lambda a_ : TORCH_DYNAMO_MODES[int(a_ )] ,default='default' ,) __UpperCamelCase : List[Any] =_ask_field( 'Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: ' ,_convert_yes_no_to_bool ,default=a_ ,error_message='Please enter yes or no.' ,) __UpperCamelCase : Union[str, Any] =_ask_field( 'Do you want to enable dynamic shape tracing? [yes/NO]: ' ,_convert_yes_no_to_bool ,default=a_ ,error_message='Please enter yes or no.' ,) __UpperCamelCase : int ='Which EC2 instance type you want to use for your training?' if distributed_type != SageMakerDistributedType.NO: __UpperCamelCase : Union[str, Any] =_ask_options( a_ ,a_ ,lambda a_ : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(a_ )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" __UpperCamelCase : List[str] =_ask_field(a_ ,lambda a_ : str(a_ ).lower() ,default='ml.p3.2xlarge' ) __UpperCamelCase : Optional[Any] =1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): __UpperCamelCase : int =_ask_field( 'How many machines do you want use? [1]: ' ,a_ ,default=1 ,) __UpperCamelCase : Optional[Any] =_ask_options( 'Do you wish to use FP16 or BF16 (mixed precision)?' ,['no', 'fp16', 'bf16', 'fp8'] ,_convert_mixed_precision ,) if use_dynamo and mixed_precision == "no": print( 'Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.' ) return SageMakerConfig( image_uri=a_ ,compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER ,distributed_type=a_ ,use_cpu=a_ ,dynamo_config=a_ ,eca_instance_type=a_ ,profile=a_ ,region=a_ ,iam_role_name=a_ ,mixed_precision=a_ ,num_machines=a_ ,sagemaker_inputs_file=a_ ,sagemaker_metrics_file=a_ ,)
71
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[str] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModel.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =AutoModel.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForPreTraining.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =AutoModelForPreTraining.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Tuple =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =TFAutoModelForCausalLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Tuple =TFAutoModelForCausalLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[int] =AutoModelForCausalLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =AutoModelForCausalLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : int =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Dict =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =TFAutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Tuple =TFAutoModelForMaskedLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[str] =AutoModelForMaskedLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : Optional[Any] =AutoModelForMaskedLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Any =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : List[Any] =TFAutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[str] =TFAutoModelForSeqaSeqLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Dict =AutoModelForSeqaSeqLM.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) __UpperCamelCase , __UpperCamelCase : List[Any] =AutoModelForSeqaSeqLM.from_pretrained( lowerCamelCase__ , output_loading_info=lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : str =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =TFAutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : str =AutoModelForSequenceClassification.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in ["bert-base-uncased"]: __UpperCamelCase : List[Any] =AutoConfig.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =TFAutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =AutoModelForQuestionAnswering.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) __UpperCamelCase : str =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =TFAutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_pt=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 ) __UpperCamelCase : int =AutoModelWithLMHead.from_pretrained(lowerCamelCase__ , from_tf=lowerCamelCase__ ) self.assertIsInstance(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(model.num_parameters() , 14410 ) self.assertEqual(model.num_parameters(only_trainable=lowerCamelCase__ ) , 14410 )
71
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { '''shi-labs/nat-mini-in1k-224''': '''https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json''', # See all Nat models at https://huggingface.co/models?filter=nat } class a__ ( a__ , a__ ): '''simple docstring''' lowercase__ : List[Any] = "nat" lowercase__ : List[str] = { "num_attention_heads": "num_heads", "num_hidden_layers": "num_layers", } def __init__( self , lowerCamelCase_=4 , lowerCamelCase_=3 , lowerCamelCase_=64 , lowerCamelCase_=[3, 4, 6, 5] , lowerCamelCase_=[2, 4, 8, 16] , lowerCamelCase_=7 , lowerCamelCase_=3.0 , lowerCamelCase_=True , lowerCamelCase_=0.0 , lowerCamelCase_=0.0 , lowerCamelCase_=0.1 , lowerCamelCase_="gelu" , lowerCamelCase_=0.02 , lowerCamelCase_=1e-5 , lowerCamelCase_=0.0 , lowerCamelCase_=None , lowerCamelCase_=None , **lowerCamelCase_ , ) -> str: super().__init__(**lowerCamelCase_ ) lowerCAmelCase__ = patch_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = embed_dim lowerCAmelCase__ = depths lowerCAmelCase__ = len(lowerCamelCase_ ) lowerCAmelCase__ = num_heads lowerCAmelCase__ = kernel_size lowerCAmelCase__ = mlp_ratio lowerCAmelCase__ = qkv_bias lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = drop_path_rate lowerCAmelCase__ = hidden_act lowerCAmelCase__ = layer_norm_eps lowerCAmelCase__ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase__ = int(embed_dim * 2 ** (len(lowerCamelCase_ ) - 1) ) lowerCAmelCase__ = layer_scale_init_value lowerCAmelCase__ = ['''stem'''] + [F"""stage{idx}""" for idx in range(1 , len(lowerCamelCase_ ) + 1 )] lowerCAmelCase__ , lowerCAmelCase__ = get_aligned_output_features_output_indices( out_features=lowerCamelCase_ , out_indices=lowerCamelCase_ , stage_names=self.stage_names )
355
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/config.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/config.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/config.json''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/config.json''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/config.json''' ), '''distilbert-base-uncased-finetuned-sst-2-english''': ( '''https://huggingface.co/distilbert-base-uncased-finetuned-sst-2-english/resolve/main/config.json''' ), } class a__ ( a__ ): '''simple docstring''' lowercase__ : Optional[Any] = "distilbert" lowercase__ : Tuple = { "hidden_size": "dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", } def __init__( self , lowerCamelCase_=3_05_22 , lowerCamelCase_=5_12 , lowerCamelCase_=False , lowerCamelCase_=6 , lowerCamelCase_=12 , lowerCamelCase_=7_68 , lowerCamelCase_=4 * 7_68 , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_="gelu" , lowerCamelCase_=0.02 , lowerCamelCase_=0.1 , lowerCamelCase_=0.2 , lowerCamelCase_=0 , **lowerCamelCase_ , ) -> Any: lowerCAmelCase__ = vocab_size lowerCAmelCase__ = max_position_embeddings lowerCAmelCase__ = sinusoidal_pos_embds lowerCAmelCase__ = n_layers lowerCAmelCase__ = n_heads lowerCAmelCase__ = dim lowerCAmelCase__ = hidden_dim lowerCAmelCase__ = dropout lowerCAmelCase__ = attention_dropout lowerCAmelCase__ = activation lowerCAmelCase__ = initializer_range lowerCAmelCase__ = qa_dropout lowerCAmelCase__ = seq_classif_dropout super().__init__(**lowerCamelCase_ , pad_token_id=lowerCamelCase_ ) class a__ ( a__ ): '''simple docstring''' @property def __SCREAMING_SNAKE_CASE ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": lowerCAmelCase__ = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: lowerCAmelCase__ = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
228
0
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _lowercase ( A__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = ConsistencyModelPipeline SCREAMING_SNAKE_CASE__ : List[Any] = UNCONDITIONAL_IMAGE_GENERATION_PARAMS SCREAMING_SNAKE_CASE__ : Optional[Any] = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt SCREAMING_SNAKE_CASE__ : List[str] = frozenset( [ '''num_inference_steps''', '''generator''', '''latents''', '''output_type''', '''return_dict''', '''callback''', '''callback_steps''', ] ) @property def __magic_name__( self :Optional[int] ) -> Optional[int]: __SCREAMING_SNAKE_CASE : Tuple = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet''' , ) return unet @property def __magic_name__( self :List[str] ) -> List[Any]: __SCREAMING_SNAKE_CASE : Dict = UNetaDModel.from_pretrained( '''diffusers/consistency-models-test''' , subfolder='''test_unet_class_cond''' , ) return unet def __magic_name__( self :Dict , lowerCAmelCase__ :List[str]=False ) -> Optional[int]: if class_cond: __SCREAMING_SNAKE_CASE : Union[str, Any] = self.dummy_cond_unet else: __SCREAMING_SNAKE_CASE : Dict = self.dummy_uncond_unet # Default to CM multistep sampler __SCREAMING_SNAKE_CASE : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __SCREAMING_SNAKE_CASE : Dict = { '''unet''': unet, '''scheduler''': scheduler, } return components def __magic_name__( self :int , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Optional[int]=0 ) -> Any: if str(lowerCAmelCase__ ).startswith('''mps''' ): __SCREAMING_SNAKE_CASE : Optional[Any] = torch.manual_seed(lowerCAmelCase__ ) else: __SCREAMING_SNAKE_CASE : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = { '''batch_size''': 1, '''num_inference_steps''': None, '''timesteps''': [22, 0], '''generator''': generator, '''output_type''': '''np''', } return inputs def __magic_name__( self :Any ) -> str: __SCREAMING_SNAKE_CASE : Union[str, Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE : List[str] = self.get_dummy_components() __SCREAMING_SNAKE_CASE : Optional[int] = ConsistencyModelPipeline(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Dict = self.get_dummy_inputs(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE : str = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : Dict = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __magic_name__( self :Optional[Any] ) -> List[str]: __SCREAMING_SNAKE_CASE : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE : Dict = self.get_dummy_components(class_cond=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = ConsistencyModelPipeline(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[Any] = self.get_dummy_inputs(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[Any] = 0 __SCREAMING_SNAKE_CASE : Tuple = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE : Dict = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : Dict = np.array([0.3572, 0.6273, 0.4031, 0.3961, 0.4321, 0.5730, 0.5266, 0.4780, 0.5004] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __magic_name__( self :str ) -> Optional[int]: __SCREAMING_SNAKE_CASE : int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE : List[str] = self.get_dummy_components() __SCREAMING_SNAKE_CASE : Optional[int] = ConsistencyModelPipeline(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[str] = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = self.get_dummy_inputs(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = 1 __SCREAMING_SNAKE_CASE : List[Any] = None __SCREAMING_SNAKE_CASE : Dict = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE : Optional[int] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : int = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def __magic_name__( self :str ) -> str: __SCREAMING_SNAKE_CASE : Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE : Any = self.get_dummy_components(class_cond=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = ConsistencyModelPipeline(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Any = self.get_dummy_inputs(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = 1 __SCREAMING_SNAKE_CASE : Union[str, Any] = None __SCREAMING_SNAKE_CASE : Union[str, Any] = 0 __SCREAMING_SNAKE_CASE : Optional[int] = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE : Dict = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : int = np.array([0.5004, 0.5004, 0.4994, 0.5008, 0.4976, 0.5018, 0.4990, 0.4982, 0.4987] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): '''simple docstring''' def __magic_name__( self :Union[str, Any] ) -> Optional[Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def __magic_name__( self :Dict , lowerCAmelCase__ :Tuple=0 , lowerCAmelCase__ :Union[str, Any]=False , lowerCAmelCase__ :int="cpu" , lowerCAmelCase__ :Optional[Any]=torch.floataa , lowerCAmelCase__ :Union[str, Any]=(1, 3, 64, 64) ) -> Optional[int]: __SCREAMING_SNAKE_CASE : List[Any] = torch.manual_seed(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = { '''num_inference_steps''': None, '''timesteps''': [22, 0], '''class_labels''': 0, '''generator''': generator, '''output_type''': '''np''', } if get_fixed_latents: __SCREAMING_SNAKE_CASE : str = self.get_fixed_latents(seed=lowerCAmelCase__ , device=lowerCAmelCase__ , dtype=lowerCAmelCase__ , shape=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Union[str, Any] = latents return inputs def __magic_name__( self :Dict , lowerCAmelCase__ :Dict=0 , lowerCAmelCase__ :List[Any]="cpu" , lowerCAmelCase__ :str=torch.floataa , lowerCAmelCase__ :Union[str, Any]=(1, 3, 64, 64) ) -> Optional[Any]: if type(lowerCAmelCase__ ) == str: __SCREAMING_SNAKE_CASE : str = torch.device(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = randn_tensor(lowerCAmelCase__ , generator=lowerCAmelCase__ , device=lowerCAmelCase__ , dtype=lowerCAmelCase__ ) return latents def __magic_name__( self :str ) -> List[str]: __SCREAMING_SNAKE_CASE : Tuple = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE : str = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __SCREAMING_SNAKE_CASE : List[str] = ConsistencyModelPipeline(unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ ) pipe.to(torch_device=lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = self.get_inputs() __SCREAMING_SNAKE_CASE : Optional[Any] = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE : Tuple = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : Dict = np.array([0.0888, 0.0881, 0.0666, 0.0479, 0.0292, 0.0195, 0.0201, 0.0163, 0.0254] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def __magic_name__( self :List[Any] ) -> Optional[int]: __SCREAMING_SNAKE_CASE : List[str] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE : Dict = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __SCREAMING_SNAKE_CASE : Optional[int] = ConsistencyModelPipeline(unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ ) pipe.to(torch_device=lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Tuple = self.get_inputs() __SCREAMING_SNAKE_CASE : str = 1 __SCREAMING_SNAKE_CASE : Optional[int] = None __SCREAMING_SNAKE_CASE : Union[str, Any] = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE : int = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : Union[str, Any] = np.array([0.0340, 0.0152, 0.0063, 0.0267, 0.0221, 0.0107, 0.0416, 0.0186, 0.0217] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 @require_torch_a def __magic_name__( self :Union[str, Any] ) -> int: __SCREAMING_SNAKE_CASE : Dict = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __SCREAMING_SNAKE_CASE : Optional[Any] = ConsistencyModelPipeline(unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ ) pipe.to(torch_device=lowerCAmelCase__ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : str = self.get_inputs(get_fixed_latents=lowerCAmelCase__ , device=lowerCAmelCase__ ) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=lowerCAmelCase__ , enable_math=lowerCAmelCase__ , enable_mem_efficient=lowerCAmelCase__ ): __SCREAMING_SNAKE_CASE : List[Any] = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE : Optional[Any] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : List[Any] = np.array([0.1875, 0.1428, 0.1289, 0.2151, 0.2092, 0.1477, 0.1877, 0.1641, 0.1353] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 @require_torch_a def __magic_name__( self :str ) -> List[Any]: __SCREAMING_SNAKE_CASE : List[str] = UNetaDModel.from_pretrained('''diffusers/consistency_models''' , subfolder='''diffusers_cd_imagenet64_l2''' ) __SCREAMING_SNAKE_CASE : Optional[int] = CMStochasticIterativeScheduler( num_train_timesteps=40 , sigma_min=0.002 , sigma_max=80.0 , ) __SCREAMING_SNAKE_CASE : Any = ConsistencyModelPipeline(unet=lowerCAmelCase__ , scheduler=lowerCAmelCase__ ) pipe.to(torch_device=lowerCAmelCase__ , torch_dtype=torch.floataa ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[Any] = self.get_inputs(get_fixed_latents=lowerCAmelCase__ , device=lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : List[Any] = 1 __SCREAMING_SNAKE_CASE : Any = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=lowerCAmelCase__ , enable_math=lowerCAmelCase__ , enable_mem_efficient=lowerCAmelCase__ ): __SCREAMING_SNAKE_CASE : List[str] = pipe(**lowerCAmelCase__ ).images assert image.shape == (1, 64, 64, 3) __SCREAMING_SNAKE_CASE : Optional[Any] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE : Any = np.array([0.1663, 0.1948, 0.2275, 0.1680, 0.1204, 0.1245, 0.1858, 0.1338, 0.2095] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3
9
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _A = '''\ ''' _A = ''' Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity ''' _A = ''' Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to \'cuda\' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = ["lorem ipsum", "Happy Birthday!", "Bienvenue"] >>> results = perplexity.compute(model_id=\'gpt2\', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 78.22 >>> print(round(results["perplexities"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric("perplexity") >>> input_texts = datasets.load_dataset("wikitext", ... "wikitext-2-raw-v1", ... split="test")["text"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=\'\'] >>> results = perplexity.compute(model_id=\'gpt2\', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) [\'perplexities\', \'mean_perplexity\'] >>> print(round(results["mean_perplexity"], 2)) 60.35 >>> print(round(results["perplexities"][0], 2)) 81.12 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ ( datasets.Metric ): def lowerCamelCase_ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """input_texts""": datasets.Value("""string""" ), } ) , reference_urls=["""https://huggingface.co/docs/transformers/perplexity"""] , ) def lowerCamelCase_ ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 1_6 , __UpperCamelCase = True , __UpperCamelCase=None ): """simple docstring""" if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": UpperCamelCase_ = """cuda""" else: UpperCamelCase_ = """cuda""" if torch.cuda.is_available() else """cpu""" UpperCamelCase_ = AutoModelForCausalLM.from_pretrained(__UpperCamelCase ) UpperCamelCase_ = model.to(__UpperCamelCase ) UpperCamelCase_ = AutoTokenizer.from_pretrained(__UpperCamelCase ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: UpperCamelCase_ = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(__UpperCamelCase ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({"""pad_token""": existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" UpperCamelCase_ = model.config.max_length - 1 else: UpperCamelCase_ = model.config.max_length UpperCamelCase_ = tokenizer( __UpperCamelCase , add_special_tokens=__UpperCamelCase , padding=__UpperCamelCase , truncation=__UpperCamelCase , max_length=__UpperCamelCase , return_tensors="""pt""" , return_attention_mask=__UpperCamelCase , ).to(__UpperCamelCase ) UpperCamelCase_ = encodings["""input_ids"""] UpperCamelCase_ = encodings["""attention_mask"""] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." UpperCamelCase_ = [] UpperCamelCase_ = CrossEntropyLoss(reduction="""none""" ) for start_index in logging.tqdm(range(0 , len(__UpperCamelCase ) , __UpperCamelCase ) ): UpperCamelCase_ = min(start_index + batch_size , len(__UpperCamelCase ) ) UpperCamelCase_ = encoded_texts[start_index:end_index] UpperCamelCase_ = attn_masks[start_index:end_index] if add_start_token: UpperCamelCase_ = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(__UpperCamelCase ) UpperCamelCase_ = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) UpperCamelCase_ = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(__UpperCamelCase ), attn_mask] , dim=1 ) UpperCamelCase_ = encoded_batch with torch.no_grad(): UpperCamelCase_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase ).logits UpperCamelCase_ = out_logits[..., :-1, :].contiguous() UpperCamelCase_ = labels[..., 1:].contiguous() UpperCamelCase_ = attn_mask[..., 1:].contiguous() UpperCamelCase_ = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , __UpperCamelCase ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(__UpperCamelCase )}
122
0
"""simple docstring""" from queue import PriorityQueue from typing import Any import numpy as np def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): for nxt, d in graph[v]: if nxt in visited_forward: continue UpperCAmelCase_ = cst_fwd.get(__lowerCamelCase , np.inf ) UpperCAmelCase_ = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) UpperCAmelCase_ = new_cost_f UpperCAmelCase_ = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: UpperCAmelCase_ = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = -1 UpperCAmelCase_ = set() UpperCAmelCase_ = set() UpperCAmelCase_ = {source: 0} UpperCAmelCase_ = {destination: 0} UpperCAmelCase_ = {source: None} UpperCAmelCase_ = {destination: None} UpperCAmelCase_ = PriorityQueue() UpperCAmelCase_ = PriorityQueue() UpperCAmelCase_ = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): UpperCAmelCase_ = queue_forward.get() visited_forward.add(__lowerCamelCase ) UpperCAmelCase_ = queue_backward.get() visited_backward.add(__lowerCamelCase ) UpperCAmelCase_ = pass_and_relaxation( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , ) UpperCAmelCase_ = pass_and_relaxation( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: UpperCAmelCase_ = shortest_distance return shortest_path_distance lowerCamelCase = { """B""": [["""C""", 1]], """C""": [["""D""", 1]], """D""": [["""F""", 1]], """E""": [["""B""", 1], ["""G""", 2]], """F""": [], """G""": [["""F""", 1]], } lowerCamelCase = { """B""": [["""E""", 1]], """C""": [["""B""", 1]], """D""": [["""C""", 1]], """F""": [["""D""", 1], ["""G""", 1]], """E""": [[None, np.inf]], """G""": [["""E""", 2]], } if __name__ == "__main__": import doctest doctest.testmod()
360
"""simple docstring""" def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): return [sentence[i : i + ngram_size] for i in range(len(lowerCAmelCase__ ) - ngram_size + 1 )] if __name__ == "__main__": from doctest import testmod testmod()
241
0
'''simple docstring''' import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE_: Optional[int] =logging.get_logger(__name__) SCREAMING_SNAKE_CASE_: Union[str, Any] ={ 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn.grep_linear': 'encoder.layers.*.attention.gru_rel_pos_linear', 'self_attn.relative_attention_bias': 'encoder.layers.*.attention.rel_attn_embed', 'self_attn.grep_a': 'encoder.layers.*.attention.gru_rel_pos_const', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } SCREAMING_SNAKE_CASE_: Optional[int] =[ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def lowerCAmelCase_ ( snake_case_ : Any , snake_case_ : List[str] , snake_case_ : str , snake_case_ : int , snake_case_ : int ) -> Optional[int]: '''simple docstring''' for attribute in key.split("." ): UpperCAmelCase_ = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if weight_type is not None: UpperCAmelCase_ = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ).shape else: UpperCAmelCase_ = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase_ = value elif weight_type == "weight_g": UpperCAmelCase_ = value elif weight_type == "weight_v": UpperCAmelCase_ = value elif weight_type == "bias": UpperCAmelCase_ = value else: UpperCAmelCase_ = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : Dict ) -> str: '''simple docstring''' UpperCAmelCase_ = [] UpperCAmelCase_ = fairseq_model.state_dict() UpperCAmelCase_ = hf_model.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase_ = False if "conv_layers" in name: load_conv_layer( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == "group" , ) UpperCAmelCase_ = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: UpperCAmelCase_ = True if "*" in mapped_key: UpperCAmelCase_ = name.split(_SCREAMING_SNAKE_CASE )[0].split("." )[-2] UpperCAmelCase_ = mapped_key.replace("*" , _SCREAMING_SNAKE_CASE ) if "weight_g" in name: UpperCAmelCase_ = "weight_g" elif "weight_v" in name: UpperCAmelCase_ = "weight_v" elif "bias" in name and "relative_attention_bias" not in name: UpperCAmelCase_ = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase_ = "weight" else: UpperCAmelCase_ = None set_recursively(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(_SCREAMING_SNAKE_CASE ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowerCAmelCase_ ( snake_case_ : Optional[Any] , snake_case_ : int , snake_case_ : List[Any] , snake_case_ : Tuple , snake_case_ : Optional[int] ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = full_name.split("conv_layers." )[-1] UpperCAmelCase_ = name.split("." ) UpperCAmelCase_ = int(items[0] ) UpperCAmelCase_ = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase_ = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase_ = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) UpperCAmelCase_ = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase_ = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_SCREAMING_SNAKE_CASE ) @torch.no_grad() def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : Optional[int] , snake_case_ : List[str]=None ) -> Tuple: '''simple docstring''' UpperCAmelCase_ = torch.load(_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ = WavLMConfigOrig(checkpoint["cfg"] ) UpperCAmelCase_ = WavLMOrig(_SCREAMING_SNAKE_CASE ) model.load_state_dict(checkpoint["model"] ) model.eval() if config_path is not None: UpperCAmelCase_ = WavLMConfig.from_pretrained(_SCREAMING_SNAKE_CASE ) else: UpperCAmelCase_ = WavLMConfig() UpperCAmelCase_ = WavLMModel(_SCREAMING_SNAKE_CASE ) recursively_load_weights(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) hf_wavlm.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_: Union[str, Any] =argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') SCREAMING_SNAKE_CASE_: Union[str, Any] =parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging SCREAMING_SNAKE_CASE__ : Optional[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ : Tuple = {'vocab_file': 'spiece.model'} SCREAMING_SNAKE_CASE__ : int = { 'vocab_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model', } } SCREAMING_SNAKE_CASE__ : str = { 'xlnet-base-cased': None, 'xlnet-large-cased': None, } # Segments (not really needed) SCREAMING_SNAKE_CASE__ : Dict = 0 SCREAMING_SNAKE_CASE__ : Tuple = 1 SCREAMING_SNAKE_CASE__ : Optional[int] = 2 SCREAMING_SNAKE_CASE__ : List[str] = 3 SCREAMING_SNAKE_CASE__ : Optional[int] = 4 class UpperCamelCase__ (lowerCAmelCase__ ): '''simple docstring''' lowerCamelCase_ : Dict = VOCAB_FILES_NAMES lowerCamelCase_ : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase_ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase_ : List[str] = """left""" def __init__( self , UpperCamelCase__ , UpperCamelCase__=False , UpperCamelCase__=True , UpperCamelCase__=False , UpperCamelCase__="<s>" , UpperCamelCase__="</s>" , UpperCamelCase__="<unk>" , UpperCamelCase__="<sep>" , UpperCamelCase__="<pad>" , UpperCamelCase__="<cls>" , UpperCamelCase__="<mask>" , UpperCamelCase__=["<eop>", "<eod>"] , UpperCamelCase__ = None , **UpperCamelCase__ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase : str = AddedToken(UpperCamelCase__ , lstrip=UpperCamelCase__ , rstrip=UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ) else mask_token lowerCamelCase : Dict = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=UpperCamelCase__ , remove_space=UpperCamelCase__ , keep_accents=UpperCamelCase__ , bos_token=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , sep_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , cls_token=UpperCamelCase__ , mask_token=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , sp_model_kwargs=self.sp_model_kwargs , **UpperCamelCase__ , ) lowerCamelCase : Any = 3 lowerCamelCase : Optional[Any] = do_lower_case lowerCamelCase : List[Any] = remove_space lowerCamelCase : str = keep_accents lowerCamelCase : List[Any] = vocab_file lowerCamelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) @property def _lowercase ( self ) -> Optional[Any]: return len(self.sp_model ) def _lowercase ( self ) -> Optional[int]: lowerCamelCase : int = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Optional[Any]: lowerCamelCase : Optional[int] = self.__dict__.copy() lowerCamelCase : Union[str, Any] = None return state def __setstate__( self , UpperCamelCase__ ) -> int: lowerCamelCase : int = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowerCamelCase : Any = {} lowerCamelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _lowercase ( self , UpperCamelCase__ ) -> Any: if self.remove_space: lowerCamelCase : Dict = " ".join(inputs.strip().split() ) else: lowerCamelCase : Union[str, Any] = inputs lowerCamelCase : Optional[Any] = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: lowerCamelCase : Optional[int] = unicodedata.normalize("NFKD" , UpperCamelCase__ ) lowerCamelCase : List[Any] = "".join([c for c in outputs if not unicodedata.combining(UpperCamelCase__ )] ) if self.do_lower_case: lowerCamelCase : List[str] = outputs.lower() return outputs def _lowercase ( self , UpperCamelCase__ ) -> List[str]: lowerCamelCase : Optional[Any] = self.preprocess_text(UpperCamelCase__ ) lowerCamelCase : Dict = self.sp_model.encode(UpperCamelCase__ , out_type=UpperCamelCase__ ) lowerCamelCase : Dict = [] for piece in pieces: if len(UpperCamelCase__ ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): lowerCamelCase : List[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(UpperCamelCase__ , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase : Union[str, Any] = cur_pieces[1:] else: lowerCamelCase : Optional[int] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(UpperCamelCase__ ) else: new_pieces.append(UpperCamelCase__ ) return new_pieces def _lowercase ( self , UpperCamelCase__ ) -> int: return self.sp_model.PieceToId(UpperCamelCase__ ) def _lowercase ( self , UpperCamelCase__ ) -> Tuple: return self.sp_model.IdToPiece(UpperCamelCase__ ) def _lowercase ( self , UpperCamelCase__ ) -> List[str]: lowerCamelCase : Union[str, Any] = "".join(UpperCamelCase__ ).replace(UpperCamelCase__ , " " ).strip() return out_string def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = False , UpperCamelCase__ = None , UpperCamelCase__ = True , **UpperCamelCase__ , ) -> str: lowerCamelCase : Optional[int] = kwargs.pop("use_source_tokenizer" , UpperCamelCase__ ) lowerCamelCase : Optional[int] = self.convert_ids_to_tokens(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCamelCase : Any = [] lowerCamelCase : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCamelCase__ ) ) lowerCamelCase : int = [] sub_texts.append(UpperCamelCase__ ) else: current_sub_text.append(UpperCamelCase__ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(UpperCamelCase__ ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens lowerCamelCase : Union[str, Any] = "".join(UpperCamelCase__ ) lowerCamelCase : Tuple = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCamelCase : int = self.clean_up_tokenization(UpperCamelCase__ ) return clean_text else: return text def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> List[int]: lowerCamelCase : str = [self.sep_token_id] lowerCamelCase : Optional[int] = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None , UpperCamelCase__ = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase__ , token_ids_a=UpperCamelCase__ , already_has_special_tokens=UpperCamelCase__ ) if token_ids_a is not None: return ([0] * len(UpperCamelCase__ )) + [1] + ([0] * len(UpperCamelCase__ )) + [1, 1] return ([0] * len(UpperCamelCase__ )) + [1, 1] def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> List[int]: lowerCamelCase : Any = [self.sep_token_id] lowerCamelCase : List[str] = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _lowercase ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> Tuple[str]: if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCamelCase : Union[str, Any] = os.path.join( UpperCamelCase__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__ , "wb" ) as fi: lowerCamelCase : str = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,)
48
0
# Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase__ : Union[str, Any] = { 'configuration_cpmant': ['CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CpmAntConfig'], 'tokenization_cpmant': ['CpmAntTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Optional[int] = [ 'CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST', 'CpmAntForCausalLM', 'CpmAntModel', 'CpmAntPreTrainedModel', ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys lowerCamelCase__ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
210
import cva import numpy as np class lowerCamelCase_ : '''simple docstring''' def __init__( self : Optional[Any] , _lowerCAmelCase : float , _lowerCAmelCase : int ): if k in (0.04, 0.06): SCREAMING_SNAKE_CASE_ = k SCREAMING_SNAKE_CASE_ = window_size else: raise ValueError('invalid k value' ) def __str__( self : Tuple ): return str(self.k ) def lowerCAmelCase_ ( self : int , _lowerCAmelCase : str ): SCREAMING_SNAKE_CASE_ = cva.imread(_lowerCAmelCase , 0 ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = img.shape SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = img.copy() SCREAMING_SNAKE_CASE_ = cva.cvtColor(_lowerCAmelCase , cva.COLOR_GRAY2RGB ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = np.gradient(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = dx**2 SCREAMING_SNAKE_CASE_ = dy**2 SCREAMING_SNAKE_CASE_ = dx * dy SCREAMING_SNAKE_CASE_ = 0.04 SCREAMING_SNAKE_CASE_ = self.window_size // 2 for y in range(_lowerCAmelCase , h - offset ): for x in range(_lowerCAmelCase , w - offset ): SCREAMING_SNAKE_CASE_ = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() SCREAMING_SNAKE_CASE_ = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() SCREAMING_SNAKE_CASE_ = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() SCREAMING_SNAKE_CASE_ = (wxx * wyy) - (wxy**2) SCREAMING_SNAKE_CASE_ = wxx + wyy SCREAMING_SNAKE_CASE_ = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": lowerCamelCase__ : Optional[int] = HarrisCorner(0.04, 3) lowerCamelCase__ , lowerCamelCase__ : str = edge_detect.detect('path_to_image') cva.imwrite('detect.png', color_img)
210
1
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = ["image_processor", "tokenizer"] lowerCamelCase__ : Dict = "BlipImageProcessor" lowerCamelCase__ : Union[str, Any] = "AutoTokenizer" def __init__( self , a , a , a ) -> Optional[int]: super().__init__(a , a ) # add QFormer tokenizer lowercase__ : Dict = qformer_tokenizer def __call__( self , a = None , a = None , a = True , a = False , a = None , a = None , a = 0 , a = None , a = None , a = False , a = False , a = False , a = False , a = False , a = True , a = None , **a , ) -> BatchFeature: if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) lowercase__ : List[Any] = BatchFeature() if text is not None: lowercase__ : Optional[int] = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) encoding.update(a ) lowercase__ : Optional[int] = self.qformer_tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) lowercase__ : List[str] = qformer_text_encoding.pop('input_ids' ) lowercase__ : Any = qformer_text_encoding.pop('attention_mask' ) if images is not None: lowercase__ : List[Any] = self.image_processor(a , return_tensors=a ) encoding.update(a ) return encoding def _UpperCAmelCase ( self , *a , **a ) -> List[str]: return self.tokenizer.batch_decode(*a , **a ) def _UpperCAmelCase ( self , *a , **a ) -> Tuple: return self.tokenizer.decode(*a , **a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : str = self.tokenizer.model_input_names lowercase__ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: if os.path.isfile(a ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(a , exist_ok=a ) lowercase__ : int = os.path.join(a , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(a ) return super().save_pretrained(a , **a ) @classmethod def _UpperCAmelCase ( cls , a , **a ) -> str: lowercase__ : str = AutoTokenizer.from_pretrained(a , subfolder='qformer_tokenizer' ) lowercase__ : int = cls._get_arguments_from_pretrained(a , **a ) args.append(a ) return cls(*a )
77
def lowercase_ (A : Union[str, Any] , A : List[str] , A : int , A : Optional[int] ): global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: snake_case__ : Union[str, Any] = mf_knapsack(i - 1 , A , A , A ) else: snake_case__ : Any = max( mf_knapsack(i - 1 , A , A , A ) , mf_knapsack(i - 1 , A , A , j - wt[i - 1] ) + val[i - 1] , ) snake_case__ : Optional[int] = val return f[i][j] def lowercase_ (A : Optional[int] , A : Union[str, Any] , A : str , A : Dict ): snake_case__ : int = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1 , n + 1 ): for w_ in range(1 , w + 1 ): if wt[i - 1] <= w_: snake_case__ : Union[str, Any] = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]] , dp[i - 1][w_] ) else: snake_case__ : str = dp[i - 1][w_] return dp[n][w_], dp def lowercase_ (A : int , A : list , A : list ): if not (isinstance(A , (list, tuple) ) and isinstance(A , (list, tuple) )): raise ValueError( 'Both the weights and values vectors must be either lists or tuples' ) snake_case__ : Dict = len(A ) if num_items != len(A ): snake_case__ : str = ( 'The number of weights must be the same as the number of values.\n' F'''But got {num_items} weights and {len(A )} values''' ) raise ValueError(A ) for i in range(A ): if not isinstance(wt[i] , A ): snake_case__ : Optional[int] = ( 'All weights must be integers but got weight of ' F'''type {type(wt[i] )} at index {i}''' ) raise TypeError(A ) snake_case__ , snake_case__ : Optional[int] = knapsack(A , A , A , A ) snake_case__ : set = set() _construct_solution(A , A , A , A , A ) return optimal_val, example_optional_set def lowercase_ (A : list , A : list , A : int , A : int , A : set ): # for the current item i at a maximum weight j to be part of an optimal subset, # the optimal value at (i, j) must be greater than the optimal value at (i-1, j). # where i - 1 means considering only the previous items at the given maximum weight if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(A , A , i - 1 , A , A ) else: optimal_set.add(A ) _construct_solution(A , A , i - 1 , j - wt[i - 1] , A ) if __name__ == "__main__": a_ :Any = [3, 2, 4, 4] a_ :List[Any] = [4, 3, 2, 3] a_ :Union[str, Any] = 4 a_ :List[str] = 6 a_ :Union[str, Any] = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] a_ , a_ :List[Any] = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 a_ , a_ :Any = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print("optimal_value = ", optimal_solution) print("An optimal subset corresponding to the optimal value", optimal_subset)
277
0
'''simple docstring''' def _a( UpperCamelCase__ : int ): '''simple docstring''' return sum(i for i in range(1, number // 2 + 1 ) if number % i == 0 ) == number if __name__ == "__main__": print('Program to check whether a number is a Perfect number or not...') a_ = int(input('Enter number: ').strip()) print(F'''{number} is {'' if perfect(number) else 'not '}a Perfect Number.''')
362
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'microsoft/biogpt': 'https://huggingface.co/microsoft/biogpt/resolve/main/config.json', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): snake_case_ = """biogpt""" def __init__( self : str , __lowercase : Union[str, Any]=4_23_84 , __lowercase : Union[str, Any]=10_24 , __lowercase : Any=24 , __lowercase : Any=16 , __lowercase : Optional[Any]=40_96 , __lowercase : Any="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : List[Any]=0.1 , __lowercase : Union[str, Any]=10_24 , __lowercase : List[Any]=0.02 , __lowercase : Tuple=1e-12 , __lowercase : Optional[Any]=True , __lowercase : Optional[Any]=True , __lowercase : Any=0.0 , __lowercase : int=0.0 , __lowercase : str=1 , __lowercase : int=0 , __lowercase : Optional[int]=2 , **__lowercase : Dict , ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ : Union[str, Any] =vocab_size SCREAMING_SNAKE_CASE__ : Optional[Any] =max_position_embeddings SCREAMING_SNAKE_CASE__ : str =hidden_size SCREAMING_SNAKE_CASE__ : int =num_hidden_layers SCREAMING_SNAKE_CASE__ : Tuple =num_attention_heads SCREAMING_SNAKE_CASE__ : Any =intermediate_size SCREAMING_SNAKE_CASE__ : int =hidden_act SCREAMING_SNAKE_CASE__ : List[Any] =hidden_dropout_prob SCREAMING_SNAKE_CASE__ : Dict =attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ : Optional[int] =initializer_range SCREAMING_SNAKE_CASE__ : Union[str, Any] =layer_norm_eps SCREAMING_SNAKE_CASE__ : Optional[Any] =scale_embedding SCREAMING_SNAKE_CASE__ : str =use_cache SCREAMING_SNAKE_CASE__ : str =layerdrop SCREAMING_SNAKE_CASE__ : Dict =activation_dropout super().__init__(pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase , **__lowercase )
222
0
import unittest import numpy as np import torch from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad class lowercase__( unittest.TestCase ): """simple docstring""" def _lowercase ( self : List[str] ) -> List[Any]: lowercase_ = 1_0 def _lowercase ( self : int ) -> List[str]: lowercase_ = [1, 2, 3, 4] lowercase_ = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] self.assertEqual(truncate_or_pad(SCREAMING_SNAKE_CASE_ , self.block_size , 0 ) , SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : int ) -> Optional[Any]: lowercase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] lowercase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] self.assertEqual(truncate_or_pad(SCREAMING_SNAKE_CASE_ , self.block_size , 0 ) , SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Union[str, Any] ) -> Optional[int]: lowercase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0, 1_1, 1_2, 1_3] lowercase_ = [1, 2, 3, 4, 5, 6, 7, 8, 9, 1_0] self.assertEqual(truncate_or_pad(SCREAMING_SNAKE_CASE_ , self.block_size , 0 ) , SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Any ) -> List[Any]: lowercase_ = '''It was the year of Our Lord one thousand seven hundred and seventy-five.\n\nSpiritual revelations were conceded to England at that favoured period, as at this.''' lowercase_ , lowercase_ = process_story(SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , [] ) def _lowercase ( self : List[str] ) -> List[str]: lowercase_ = '''''' lowercase_ , lowercase_ = process_story(SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , [] ) self.assertEqual(SCREAMING_SNAKE_CASE_ , [] ) def _lowercase ( self : Union[str, Any] ) -> Union[str, Any]: lowercase_ = ( '''It was the year of Our Lord one thousand seven hundred and ''' '''seventy-five\n\nSpiritual revelations were conceded to England ''' '''at that favoured period, as at this.\n@highlight\n\nIt was the best of times''' ) lowercase_ , lowercase_ = process_story(SCREAMING_SNAKE_CASE_ ) lowercase_ = [ '''It was the year of Our Lord one thousand seven hundred and seventy-five.''', '''Spiritual revelations were conceded to England at that favoured period, as at this.''', ] self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) lowercase_ = ['''It was the best of times.'''] self.assertEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) def _lowercase ( self : Union[str, Any] ) -> Optional[Any]: lowercase_ = torch.tensor([1, 2, 3, 4] ) lowercase_ = torch.tensor([1, 1, 1, 1] ) np.testing.assert_array_equal(build_mask(SCREAMING_SNAKE_CASE_ , 0 ).numpy() , expected.numpy() ) def _lowercase ( self : List[Any] ) -> Tuple: lowercase_ = torch.tensor([1, 2, 3, 4, 2_3, 2_3, 2_3] ) lowercase_ = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(SCREAMING_SNAKE_CASE_ , 2_3 ).numpy() , expected.numpy() ) def _lowercase ( self : int ) -> Dict: lowercase_ = torch.tensor([8, 2, 3, 4, 1, 1, 1] ) lowercase_ = torch.tensor([1, 1, 1, 1, 0, 0, 0] ) np.testing.assert_array_equal(build_mask(SCREAMING_SNAKE_CASE_ , 1 ).numpy() , expected.numpy() ) def _lowercase ( self : List[str] ) -> Tuple: lowercase_ = 1_0_1 lowercase_ = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 1_0_1, 5, 6], [1, 1_0_1, 3, 4, 1_0_1, 6]] ) lowercase_ = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]] ) lowercase_ = compute_token_type_ids(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) np.testing.assert_array_equal(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )
30
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCamelCase = { '''configuration_clap''': [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapAudioConfig''', '''ClapConfig''', '''ClapTextConfig''', ], '''processing_clap''': ['''ClapProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapModel''', '''ClapPreTrainedModel''', '''ClapTextModel''', '''ClapTextModelWithProjection''', '''ClapAudioModel''', '''ClapAudioModelWithProjection''', ] __lowerCamelCase = ['''ClapFeatureExtractor'''] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys __lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
162
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { '''configuration_clap''': [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapAudioConfig''', '''ClapConfig''', '''ClapTextConfig''', ], '''processing_clap''': ['''ClapProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapModel''', '''ClapPreTrainedModel''', '''ClapTextModel''', '''ClapTextModelWithProjection''', '''ClapAudioModel''', '''ClapAudioModelWithProjection''', ] _snake_case = ['''ClapFeatureExtractor'''] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys _snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
367
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _snake_case = logging.get_logger(__name__) _snake_case = { '''facebook/detr-resnet-50''': '''https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json''', # See all DETR models at https://huggingface.co/models?filter=detr } class _snake_case ( _lowercase ): lowerCamelCase__: str = "detr" lowerCamelCase__: Dict = ["past_key_values"] lowerCamelCase__: str = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self: List[str] , __lowerCamelCase: List[Any]=True , __lowerCamelCase: Any=None , __lowerCamelCase: Dict=3 , __lowerCamelCase: str=1_00 , __lowerCamelCase: Union[str, Any]=6 , __lowerCamelCase: Union[str, Any]=20_48 , __lowerCamelCase: Dict=8 , __lowerCamelCase: Optional[int]=6 , __lowerCamelCase: List[Any]=20_48 , __lowerCamelCase: int=8 , __lowerCamelCase: Tuple=0.0 , __lowerCamelCase: Dict=0.0 , __lowerCamelCase: Any=True , __lowerCamelCase: Tuple="relu" , __lowerCamelCase: Tuple=2_56 , __lowerCamelCase: Dict=0.1 , __lowerCamelCase: Union[str, Any]=0.0 , __lowerCamelCase: Optional[int]=0.0 , __lowerCamelCase: Union[str, Any]=0.02 , __lowerCamelCase: str=1.0 , __lowerCamelCase: List[str]=False , __lowerCamelCase: Dict="sine" , __lowerCamelCase: Optional[int]="resnet50" , __lowerCamelCase: Optional[int]=True , __lowerCamelCase: int=False , __lowerCamelCase: Union[str, Any]=1 , __lowerCamelCase: Tuple=5 , __lowerCamelCase: int=2 , __lowerCamelCase: Dict=1 , __lowerCamelCase: Dict=1 , __lowerCamelCase: Union[str, Any]=5 , __lowerCamelCase: Dict=2 , __lowerCamelCase: int=0.1 , **__lowerCamelCase: str , ) -> int: if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) __UpperCAmelCase : Optional[int] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(__lowerCamelCase , __lowerCamelCase ): __UpperCAmelCase : List[Any] = backbone_config.get("model_type" ) __UpperCAmelCase : List[str] = CONFIG_MAPPING[backbone_model_type] __UpperCAmelCase : List[str] = config_class.from_dict(__lowerCamelCase ) # set timm attributes to None __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase : List[Any] = None, None, None __UpperCAmelCase : Any = use_timm_backbone __UpperCAmelCase : Optional[Any] = backbone_config __UpperCAmelCase : Optional[Any] = num_channels __UpperCAmelCase : List[Any] = num_queries __UpperCAmelCase : Optional[int] = d_model __UpperCAmelCase : Optional[Any] = encoder_ffn_dim __UpperCAmelCase : Dict = encoder_layers __UpperCAmelCase : List[Any] = encoder_attention_heads __UpperCAmelCase : int = decoder_ffn_dim __UpperCAmelCase : Tuple = decoder_layers __UpperCAmelCase : int = decoder_attention_heads __UpperCAmelCase : List[Any] = dropout __UpperCAmelCase : Dict = attention_dropout __UpperCAmelCase : Optional[Any] = activation_dropout __UpperCAmelCase : int = activation_function __UpperCAmelCase : Any = init_std __UpperCAmelCase : str = init_xavier_std __UpperCAmelCase : int = encoder_layerdrop __UpperCAmelCase : Tuple = decoder_layerdrop __UpperCAmelCase : List[Any] = encoder_layers __UpperCAmelCase : Optional[Any] = auxiliary_loss __UpperCAmelCase : int = position_embedding_type __UpperCAmelCase : Optional[int] = backbone __UpperCAmelCase : str = use_pretrained_backbone __UpperCAmelCase : Dict = dilation # Hungarian matcher __UpperCAmelCase : Optional[int] = class_cost __UpperCAmelCase : Optional[Any] = bbox_cost __UpperCAmelCase : Optional[int] = giou_cost # Loss coefficients __UpperCAmelCase : Any = mask_loss_coefficient __UpperCAmelCase : Any = dice_loss_coefficient __UpperCAmelCase : Any = bbox_loss_coefficient __UpperCAmelCase : Optional[int] = giou_loss_coefficient __UpperCAmelCase : Optional[Any] = eos_coefficient super().__init__(is_encoder_decoder=__lowerCamelCase , **__lowerCamelCase ) @property def _lowerCamelCase ( self: Dict ) -> int: return self.encoder_attention_heads @property def _lowerCamelCase ( self: str ) -> int: return self.d_model @classmethod def _lowerCamelCase ( cls: Optional[int] , __lowerCamelCase: PretrainedConfig , **__lowerCamelCase: List[Any] ) -> List[Any]: return cls(backbone_config=__lowerCamelCase , **__lowerCamelCase ) def _lowerCamelCase ( self: str ) -> Dict[str, any]: __UpperCAmelCase : Optional[int] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __UpperCAmelCase : int = self.backbone_config.to_dict() __UpperCAmelCase : List[str] = self.__class__.model_type return output class _snake_case ( _lowercase ): lowerCamelCase__: Optional[int] = version.parse("1.11" ) @property def _lowerCamelCase ( self: Optional[Any] ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def _lowerCamelCase ( self: Optional[Any] ) -> float: return 1e-5 @property def _lowerCamelCase ( self: List[str] ) -> int: return 12
342
0
"""simple docstring""" # Function to print upper half of diamond (pyramid) def __UpperCAmelCase ( lowercase ): """simple docstring""" for i in range(0 ,_lowerCAmelCase ): for _ in range(0 ,n - i - 1 ): # printing spaces print(""" """ ,end="""""" ) for _ in range(0 ,i + 1 ): # printing stars print("""* """ ,end="""""" ) print() def __UpperCAmelCase ( lowercase ): """simple docstring""" for i in range(_lowerCAmelCase ,0 ,-1 ): for _ in range(_lowerCAmelCase ,0 ,-1 ): # printing stars print("""* """ ,end="""""" ) print() for _ in range(n - i + 1 ,0 ,-1 ): # printing spaces print(""" """ ,end="""""" ) def __UpperCAmelCase ( lowercase ): """simple docstring""" if n <= 0: print(""" ... .... nothing printing :(""" ) return floyd(_lowerCAmelCase ) # upper half reverse_floyd(_lowerCAmelCase ) # lower half if __name__ == "__main__": print(r"""| /\ | |- | |- |--| |\ /| |-""") print(r"""|/ \| |- |_ |_ |__| | \/ | |_""") UpperCAmelCase__ = 1 while K: UpperCAmelCase__ = int(input("""enter the number and , and see the magic : """)) print() pretty_print(user_number) UpperCAmelCase__ = int(input("""press 0 to exit... and 1 to continue...""")) print("""Good Bye...""")
289
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : List[Any] = logging.get_logger(__name__) __lowerCamelCase : str = { """roberta-base""": """https://huggingface.co/roberta-base/resolve/main/config.json""", """roberta-large""": """https://huggingface.co/roberta-large/resolve/main/config.json""", """roberta-large-mnli""": """https://huggingface.co/roberta-large-mnli/resolve/main/config.json""", """distilroberta-base""": """https://huggingface.co/distilroberta-base/resolve/main/config.json""", """roberta-base-openai-detector""": """https://huggingface.co/roberta-base-openai-detector/resolve/main/config.json""", """roberta-large-openai-detector""": """https://huggingface.co/roberta-large-openai-detector/resolve/main/config.json""", } class A__ ( __snake_case ): _UpperCAmelCase :Union[str, Any] = 'roberta' def __init__( self , A_=5_0265 , A_=768 , A_=12 , A_=12 , A_=3072 , A_="gelu" , A_=0.1 , A_=0.1 , A_=512 , A_=2 , A_=0.02 , A_=1e-12 , A_=1 , A_=0 , A_=2 , A_="absolute" , A_=True , A_=None , **A_ , ): '''simple docstring''' super().__init__(pad_token_id=A_ , bos_token_id=A_ , eos_token_id=A_ , **A_ ) UpperCamelCase : Optional[int] = vocab_size UpperCamelCase : Dict = hidden_size UpperCamelCase : str = num_hidden_layers UpperCamelCase : Any = num_attention_heads UpperCamelCase : List[str] = hidden_act UpperCamelCase : Optional[Any] = intermediate_size UpperCamelCase : Tuple = hidden_dropout_prob UpperCamelCase : Tuple = attention_probs_dropout_prob UpperCamelCase : Tuple = max_position_embeddings UpperCamelCase : Any = type_vocab_size UpperCamelCase : int = initializer_range UpperCamelCase : str = layer_norm_eps UpperCamelCase : Dict = position_embedding_type UpperCamelCase : Any = use_cache UpperCamelCase : Union[str, Any] = classifier_dropout class A__ ( __snake_case ): @property def __UpperCamelCase( self ): '''simple docstring''' if self.task == "multiple-choice": UpperCamelCase : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCamelCase : Optional[int] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
52
0
"""simple docstring""" from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig a_ = { 'susnato/ernie-m-base_pytorch': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json', 'susnato/ernie-m-large_pytorch': 'https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json', } class __snake_case ( _UpperCamelCase ): """simple docstring""" _lowerCamelCase = 'ernie_m' _lowerCamelCase = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self , __lowerCamelCase = 25_0002 , __lowerCamelCase = 768 , __lowerCamelCase = 12 , __lowerCamelCase = 12 , __lowerCamelCase = 3072 , __lowerCamelCase = "gelu" , __lowerCamelCase = 0.1 , __lowerCamelCase = 0.1 , __lowerCamelCase = 514 , __lowerCamelCase = 0.0_2 , __lowerCamelCase = 1 , __lowerCamelCase = 1e-0_5 , __lowerCamelCase=None , __lowerCamelCase=False , __lowerCamelCase=0.0 , **__lowerCamelCase , ): '''simple docstring''' super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) __A : Optional[int] = vocab_size __A : Optional[int] = hidden_size __A : str = num_hidden_layers __A : Tuple = num_attention_heads __A : int = intermediate_size __A : List[Any] = hidden_act __A : Any = hidden_dropout_prob __A : int = attention_probs_dropout_prob __A : Dict = max_position_embeddings __A : List[str] = initializer_range __A : List[Any] = layer_norm_eps __A : int = classifier_dropout __A : Dict = is_decoder __A : List[Any] = act_dropout
368
"""simple docstring""" import numpy as np import qiskit def __lowercase ( snake_case_ : int = 8 ,snake_case_ : int | None = None ) ->str: '''simple docstring''' __A : str = np.random.default_rng(seed=snake_case_ ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. __A : str = 6 * key_len # Measurement basis for Alice's qubits. __A : Any = rng.integers(2 ,size=snake_case_ ) # The set of states Alice will prepare. __A : Any = rng.integers(2 ,size=snake_case_ ) # Measurement basis for Bob's qubits. __A : str = rng.integers(2 ,size=snake_case_ ) # Quantum Circuit to simulate BB84 __A : Dict = qiskit.QuantumCircuit(snake_case_ ,name='''BB84''' ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(snake_case_ ): if alice_state[index] == 1: bbaa_circ.x(snake_case_ ) if alice_basis[index] == 1: bbaa_circ.h(snake_case_ ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(snake_case_ ): if bob_basis[index] == 1: bbaa_circ.h(snake_case_ ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. __A : List[str] = qiskit.Aer.get_backend('''aer_simulator''' ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. __A : List[str] = qiskit.execute(snake_case_ ,snake_case_ ,shots=1 ,seed_simulator=snake_case_ ) # Returns the result of measurement. __A : Union[str, Any] = job.result().get_counts(snake_case_ ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. __A : int = ''''''.join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( snake_case_ ,snake_case_ ,snake_case_ ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. __A : Union[str, Any] = gen_key[:key_len] if len(snake_case_ ) >= key_len else gen_key.ljust(snake_case_ ,'''0''' ) return key if __name__ == "__main__": print(f'''The generated key is : {bbaa(8, seed=0)}''') from doctest import testmod testmod()
291
0
"""simple docstring""" from __future__ import annotations def snake_case__ ( __lowerCamelCase : int , __lowerCamelCase : int ): """simple docstring""" lowerCamelCase__ : list[list[int]] =[] create_all_state(1 , __lowerCamelCase , __lowerCamelCase , [] , __lowerCamelCase ) return result def snake_case__ ( __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : list[int] , __lowerCamelCase : list[list[int]] , ): """simple docstring""" if level == 0: total_list.append(current_list[:] ) return for i in range(__lowerCamelCase , total_number - level + 2 ): current_list.append(__lowerCamelCase ) create_all_state(i + 1 , __lowerCamelCase , level - 1 , __lowerCamelCase , __lowerCamelCase ) current_list.pop() def snake_case__ ( __lowerCamelCase : list[list[int]] ): """simple docstring""" for i in total_list: print(*__lowerCamelCase ) if __name__ == "__main__": _lowercase : List[Any] = 4 _lowercase : Union[str, Any] = 2 _lowercase : Any = generate_all_combinations(n, k) print_all_state(total_list)
238
"""simple docstring""" from __future__ import annotations _lowercase : Dict = 1.6_021E-19 # units = C def snake_case__ ( __lowerCamelCase : float , __lowerCamelCase : float , __lowerCamelCase : float , ): """simple docstring""" if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif conductivity < 0: raise ValueError('''Conductivity cannot be negative''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative''' ) elif mobility < 0: raise ValueError('''mobility cannot be negative''' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
238
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class _lowerCAmelCase ( a ): """simple docstring""" __magic_name__ :Dict = """audio-spectrogram-transformer""" def __init__( self , __UpperCAmelCase=7_6_8 , __UpperCAmelCase=1_2 , __UpperCAmelCase=1_2 , __UpperCAmelCase=3_0_7_2 , __UpperCAmelCase="gelu" , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-12 , __UpperCAmelCase=1_6 , __UpperCAmelCase=True , __UpperCAmelCase=1_0 , __UpperCAmelCase=1_0 , __UpperCAmelCase=1_0_2_4 , __UpperCAmelCase=1_2_8 , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(**__UpperCAmelCase ) lowerCAmelCase__ :Any = hidden_size lowerCAmelCase__ :Optional[int] = num_hidden_layers lowerCAmelCase__ :Any = num_attention_heads lowerCAmelCase__ :Optional[Any] = intermediate_size lowerCAmelCase__ :Tuple = hidden_act lowerCAmelCase__ :List[Any] = hidden_dropout_prob lowerCAmelCase__ :Dict = attention_probs_dropout_prob lowerCAmelCase__ :List[str] = initializer_range lowerCAmelCase__ :Dict = layer_norm_eps lowerCAmelCase__ :Any = patch_size lowerCAmelCase__ :Optional[int] = qkv_bias lowerCAmelCase__ :Any = frequency_stride lowerCAmelCase__ :int = time_stride lowerCAmelCase__ :Optional[int] = max_length lowerCAmelCase__ :List[str] = num_mel_bins
369
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices __A = logging.get_logger(__name__) __A = { """microsoft/swin-tiny-patch4-window7-224""": ( """https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json""" ), # See all Swin models at https://huggingface.co/models?filter=swin } class _lowerCAmelCase ( a , a ): """simple docstring""" __magic_name__ :int = """swin""" __magic_name__ :Tuple = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , __UpperCAmelCase=2_2_4 , __UpperCAmelCase=4 , __UpperCAmelCase=3 , __UpperCAmelCase=9_6 , __UpperCAmelCase=[2, 2, 6, 2] , __UpperCAmelCase=[3, 6, 1_2, 2_4] , __UpperCAmelCase=7 , __UpperCAmelCase=4.0 , __UpperCAmelCase=True , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.1 , __UpperCAmelCase="gelu" , __UpperCAmelCase=False , __UpperCAmelCase=0.02 , __UpperCAmelCase=1E-5 , __UpperCAmelCase=3_2 , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase , ): '''simple docstring''' super().__init__(**__UpperCAmelCase ) lowerCAmelCase__ :Any = image_size lowerCAmelCase__ :List[Any] = patch_size lowerCAmelCase__ :Optional[int] = num_channels lowerCAmelCase__ :str = embed_dim lowerCAmelCase__ :Optional[int] = depths lowerCAmelCase__ :List[str] = len(__UpperCAmelCase ) lowerCAmelCase__ :Optional[Any] = num_heads lowerCAmelCase__ :List[Any] = window_size lowerCAmelCase__ :List[Any] = mlp_ratio lowerCAmelCase__ :int = qkv_bias lowerCAmelCase__ :Optional[int] = hidden_dropout_prob lowerCAmelCase__ :int = attention_probs_dropout_prob lowerCAmelCase__ :List[Any] = drop_path_rate lowerCAmelCase__ :Any = hidden_act lowerCAmelCase__ :Dict = use_absolute_embeddings lowerCAmelCase__ :int = layer_norm_eps lowerCAmelCase__ :Dict = initializer_range lowerCAmelCase__ :int = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase__ :str = int(embed_dim * 2 ** (len(__UpperCAmelCase ) - 1) ) lowerCAmelCase__ :str = ['stem'] + [F"stage{idx}" for idx in range(1 , len(__UpperCAmelCase ) + 1 )] lowerCAmelCase__ , lowerCAmelCase__ :List[Any] = get_aligned_output_features_output_indices( out_features=__UpperCAmelCase , out_indices=__UpperCAmelCase , stage_names=self.stage_names ) class _lowerCAmelCase ( a ): """simple docstring""" __magic_name__ :int = version.parse("""1.11""" ) @property def snake_case ( self ): '''simple docstring''' return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def snake_case ( self ): '''simple docstring''' return 1E-4
254
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __A ={ '''configuration_mobilebert''': [ '''MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileBertConfig''', '''MobileBertOnnxConfig''', ], '''tokenization_mobilebert''': ['''MobileBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =['''MobileBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileBertForMaskedLM''', '''MobileBertForMultipleChoice''', '''MobileBertForNextSentencePrediction''', '''MobileBertForPreTraining''', '''MobileBertForQuestionAnswering''', '''MobileBertForSequenceClassification''', '''MobileBertForTokenClassification''', '''MobileBertLayer''', '''MobileBertModel''', '''MobileBertPreTrainedModel''', '''load_tf_weights_in_mobilebert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFMobileBertForMaskedLM''', '''TFMobileBertForMultipleChoice''', '''TFMobileBertForNextSentencePrediction''', '''TFMobileBertForPreTraining''', '''TFMobileBertForQuestionAnswering''', '''TFMobileBertForSequenceClassification''', '''TFMobileBertForTokenClassification''', '''TFMobileBertMainLayer''', '''TFMobileBertModel''', '''TFMobileBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys __A =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
19
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A ={'''configuration_wavlm''': ['''WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''WavLMConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''WavLMForAudioFrameClassification''', '''WavLMForCTC''', '''WavLMForSequenceClassification''', '''WavLMForXVector''', '''WavLMModel''', '''WavLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavlm import WAVLM_PRETRAINED_CONFIG_ARCHIVE_MAP, WavLMConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavlm import ( WAVLM_PRETRAINED_MODEL_ARCHIVE_LIST, WavLMForAudioFrameClassification, WavLMForCTC, WavLMForSequenceClassification, WavLMForXVector, WavLMModel, WavLMPreTrainedModel, ) else: import sys __A =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
19
1
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging _A = logging.get_logger(__name__) _A = { 'speechbrain/m-ctc-t-large': 'https://huggingface.co/speechbrain/m-ctc-t-large/resolve/main/config.json', # See all M-CTC-T models at https://huggingface.co/models?filter=mctct } class _lowercase ( __UpperCAmelCase ): lowercase_ = 'mctct' def __init__( self , UpperCAmelCase_=8065 , UpperCAmelCase_=1536 , UpperCAmelCase_=36 , UpperCAmelCase_=6144 , UpperCAmelCase_=4 , UpperCAmelCase_=384 , UpperCAmelCase_=920 , UpperCAmelCase_=1E-5 , UpperCAmelCase_=0.3 , UpperCAmelCase_="relu" , UpperCAmelCase_=0.02 , UpperCAmelCase_=0.3 , UpperCAmelCase_=0.3 , UpperCAmelCase_=1 , UpperCAmelCase_=0 , UpperCAmelCase_=2 , UpperCAmelCase_=1 , UpperCAmelCase_=0.3 , UpperCAmelCase_=1 , UpperCAmelCase_=(7,) , UpperCAmelCase_=(3,) , UpperCAmelCase_=80 , UpperCAmelCase_=1 , UpperCAmelCase_=None , UpperCAmelCase_="sum" , UpperCAmelCase_=False , **UpperCAmelCase_ , ) -> int: super().__init__(**UpperCAmelCase_ , pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ ) lowerCamelCase : str = vocab_size lowerCamelCase : List[Any] = hidden_size lowerCamelCase : Dict = num_hidden_layers lowerCamelCase : Optional[int] = intermediate_size lowerCamelCase : Tuple = num_attention_heads lowerCamelCase : Dict = attention_head_dim lowerCamelCase : str = max_position_embeddings lowerCamelCase : Optional[int] = layer_norm_eps lowerCamelCase : Dict = layerdrop lowerCamelCase : Any = hidden_act lowerCamelCase : List[Any] = initializer_range lowerCamelCase : List[Any] = hidden_dropout_prob lowerCamelCase : Any = attention_probs_dropout_prob lowerCamelCase : Any = pad_token_id lowerCamelCase : List[Any] = bos_token_id lowerCamelCase : Dict = eos_token_id lowerCamelCase : Any = conv_glu_dim lowerCamelCase : str = conv_dropout lowerCamelCase : Union[str, Any] = num_conv_layers lowerCamelCase : Tuple = input_feat_per_channel lowerCamelCase : List[str] = input_channels lowerCamelCase : str = conv_channels lowerCamelCase : Any = ctc_loss_reduction lowerCamelCase : Optional[Any] = ctc_zero_infinity # prevents config testing fail with exporting to json lowerCamelCase : int = list(UpperCAmelCase_ ) lowerCamelCase : str = list(UpperCAmelCase_ ) if len(self.conv_kernel ) != self.num_conv_layers: raise ValueError( 'Configuration for convolutional module is incorrect. ' 'It is required that `len(config.conv_kernel)` == `config.num_conv_layers` ' F"""but is `len(config.conv_kernel) = {len(self.conv_kernel )}`, """ F"""`config.num_conv_layers = {self.num_conv_layers}`.""" )
205
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import requests import torch from PIL import Image from transformers import GLPNConfig, GLPNForDepthEstimation, GLPNImageProcessor from transformers.utils import logging logging.set_verbosity_info() _A = logging.get_logger(__name__) def UpperCAmelCase ( a_ ): '''simple docstring''' lowerCamelCase : int = OrderedDict() for key, value in state_dict.items(): if key.startswith('module.encoder' ): lowerCamelCase : Tuple = key.replace('module.encoder', 'glpn.encoder' ) if key.startswith('module.decoder' ): lowerCamelCase : str = key.replace('module.decoder', 'decoder.stages' ) if "patch_embed" in key: # replace for example patch_embed1 by patch_embeddings.0 lowerCamelCase : Any = key[key.find('patch_embed' ) + len('patch_embed' )] lowerCamelCase : Dict = key.replace(F"""patch_embed{idx}""", F"""patch_embeddings.{int(a_ )-1}""" ) if "norm" in key: lowerCamelCase : Optional[int] = key.replace('norm', 'layer_norm' ) if "glpn.encoder.layer_norm" in key: # replace for example layer_norm1 by layer_norm.0 lowerCamelCase : List[str] = key[key.find('glpn.encoder.layer_norm' ) + len('glpn.encoder.layer_norm' )] lowerCamelCase : List[str] = key.replace(F"""layer_norm{idx}""", F"""layer_norm.{int(a_ )-1}""" ) if "layer_norm1" in key: lowerCamelCase : List[Any] = key.replace('layer_norm1', 'layer_norm_1' ) if "layer_norm2" in key: lowerCamelCase : str = key.replace('layer_norm2', 'layer_norm_2' ) if "block" in key: # replace for example block1 by block.0 lowerCamelCase : Union[str, Any] = key[key.find('block' ) + len('block' )] lowerCamelCase : List[str] = key.replace(F"""block{idx}""", F"""block.{int(a_ )-1}""" ) if "attn.q" in key: lowerCamelCase : Union[str, Any] = key.replace('attn.q', 'attention.self.query' ) if "attn.proj" in key: lowerCamelCase : Dict = key.replace('attn.proj', 'attention.output.dense' ) if "attn" in key: lowerCamelCase : int = key.replace('attn', 'attention.self' ) if "fc1" in key: lowerCamelCase : Any = key.replace('fc1', 'dense1' ) if "fc2" in key: lowerCamelCase : List[Any] = key.replace('fc2', 'dense2' ) if "linear_pred" in key: lowerCamelCase : Optional[Any] = key.replace('linear_pred', 'classifier' ) if "linear_fuse" in key: lowerCamelCase : Union[str, Any] = key.replace('linear_fuse.conv', 'linear_fuse' ) lowerCamelCase : Optional[int] = key.replace('linear_fuse.bn', 'batch_norm' ) if "linear_c" in key: # replace for example linear_c4 by linear_c.3 lowerCamelCase : str = key[key.find('linear_c' ) + len('linear_c' )] lowerCamelCase : List[Any] = key.replace(F"""linear_c{idx}""", F"""linear_c.{int(a_ )-1}""" ) if "bot_conv" in key: lowerCamelCase : int = key.replace('bot_conv', '0.convolution' ) if "skip_conv1" in key: lowerCamelCase : Any = key.replace('skip_conv1', '1.convolution' ) if "skip_conv2" in key: lowerCamelCase : Optional[Any] = key.replace('skip_conv2', '2.convolution' ) if "fusion1" in key: lowerCamelCase : str = key.replace('fusion1', '1.fusion' ) if "fusion2" in key: lowerCamelCase : Optional[Any] = key.replace('fusion2', '2.fusion' ) if "fusion3" in key: lowerCamelCase : List[str] = key.replace('fusion3', '3.fusion' ) if "fusion" in key and "conv" in key: lowerCamelCase : Optional[int] = key.replace('conv', 'convolutional_layer' ) if key.startswith('module.last_layer_depth' ): lowerCamelCase : Tuple = key.replace('module.last_layer_depth', 'head.head' ) lowerCamelCase : List[Any] = value return new_state_dict def UpperCAmelCase ( a_, a_ ): '''simple docstring''' for i in range(config.num_encoder_blocks ): for j in range(config.depths[i] ): # read in weights + bias of keys and values (which is a single matrix in the original implementation) lowerCamelCase : Any = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.weight""" ) lowerCamelCase : Optional[Any] = state_dict.pop(F"""glpn.encoder.block.{i}.{j}.attention.self.kv.bias""" ) # next, add keys and values (in that order) to the state dict lowerCamelCase : Any = kv_weight[ : config.hidden_sizes[i], : ] lowerCamelCase : List[Any] = kv_bias[: config.hidden_sizes[i]] lowerCamelCase : Dict = kv_weight[ config.hidden_sizes[i] :, : ] lowerCamelCase : List[Any] = kv_bias[config.hidden_sizes[i] :] def UpperCAmelCase ( ): '''simple docstring''' lowerCamelCase : Dict = 'http://images.cocodataset.org/val2017/000000039769.jpg' lowerCamelCase : List[Any] = Image.open(requests.get(a_, stream=a_ ).raw ) return image @torch.no_grad() def UpperCAmelCase ( a_, a_, a_=False, a_=None ): '''simple docstring''' lowerCamelCase : int = GLPNConfig(hidden_sizes=[64, 128, 320, 512], decoder_hidden_size=64, depths=[3, 8, 27, 3] ) # load image processor (only resize + rescale) lowerCamelCase : Any = GLPNImageProcessor() # prepare image lowerCamelCase : int = prepare_img() lowerCamelCase : Tuple = image_processor(images=a_, return_tensors='pt' ).pixel_values logger.info('Converting model...' ) # load original state dict lowerCamelCase : Optional[Any] = torch.load(a_, map_location=torch.device('cpu' ) ) # rename keys lowerCamelCase : Any = rename_keys(a_ ) # key and value matrices need special treatment read_in_k_v(a_, a_ ) # create HuggingFace model and load state dict lowerCamelCase : Optional[int] = GLPNForDepthEstimation(a_ ) model.load_state_dict(a_ ) model.eval() # forward pass lowerCamelCase : str = model(a_ ) lowerCamelCase : Any = outputs.predicted_depth # verify output if model_name is not None: if "nyu" in model_name: lowerCamelCase : Any = torch.tensor( [[4.4_1_4_7, 4.0_8_7_3, 4.0_6_7_3], [3.7_8_9_0, 3.2_8_8_1, 3.1_5_2_5], [3.7_6_7_4, 3.5_4_2_3, 3.4_9_1_3]] ) elif "kitti" in model_name: lowerCamelCase : str = torch.tensor( [[3.4_2_9_1, 2.7_8_6_5, 2.5_1_5_1], [3.2_8_4_1, 2.7_0_2_1, 2.3_5_0_2], [3.1_1_4_7, 2.4_6_2_5, 2.2_4_8_1]] ) else: raise ValueError(F"""Unknown model name: {model_name}""" ) lowerCamelCase : int = torch.Size([1, 480, 640] ) assert predicted_depth.shape == expected_shape assert torch.allclose(predicted_depth[0, :3, :3], a_, atol=1E-4 ) print('Looks ok!' ) # finally, push to hub if required if push_to_hub: logger.info('Pushing model and image processor to the hub...' ) model.push_to_hub( repo_path_or_name=Path(a_, a_ ), organization='nielsr', commit_message='Add model', use_temp_dir=a_, ) image_processor.push_to_hub( repo_path_or_name=Path(a_, a_ ), organization='nielsr', commit_message='Add image processor', use_temp_dir=a_, ) if __name__ == "__main__": _A = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to upload the model to the HuggingFace hub.' ) parser.add_argument( '--model_name', default='glpn-kitti', type=str, help='Name of the model in case you\'re pushing to the hub.', ) _A = parser.parse_args() convert_glpn_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
205
1
'''simple docstring''' import math from enum import Enum from typing import Optional, Union from torch.optim import Optimizer from torch.optim.lr_scheduler import LambdaLR from .utils import logging A_ : str = logging.get_logger(__name__) class lowercase ( _lowerCamelCase ): """simple docstring""" UpperCAmelCase = """linear""" UpperCAmelCase = """cosine""" UpperCAmelCase = """cosine_with_restarts""" UpperCAmelCase = """polynomial""" UpperCAmelCase = """constant""" UpperCAmelCase = """constant_with_warmup""" UpperCAmelCase = """piecewise_constant""" def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ = -1 )-> List[Any]: '''simple docstring''' return LambdaLR(__lowerCamelCase , lambda lowerCAmelCase_ : 1 , last_epoch=__lowerCamelCase ) def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = -1 )-> Tuple: '''simple docstring''' def lr_lambda(lowerCAmelCase_ ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1.0 , __lowerCamelCase ) ) return 1.0 return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = -1 )-> str: '''simple docstring''' _UpperCAmelCase : List[Any] = {} _UpperCAmelCase : Any = step_rules.split(""",""" ) for rule_str in rule_list[:-1]: _UpperCAmelCase ,_UpperCAmelCase : Dict = rule_str.split(""":""" ) _UpperCAmelCase : List[Any] = int(__lowerCamelCase ) _UpperCAmelCase : Optional[Any] = float(__lowerCamelCase ) _UpperCAmelCase : Any = value _UpperCAmelCase : str = float(rule_list[-1] ) def create_rules_function(lowerCAmelCase_ , lowerCAmelCase_ ): def rule_func(lowerCAmelCase_ ) -> float: _UpperCAmelCase : Optional[Any] = sorted(rules_dict.keys() ) for i, sorted_step in enumerate(__lowerCamelCase ): if steps < sorted_step: return rules_dict[sorted_steps[i]] return last_lr_multiple return rule_func _UpperCAmelCase : str = create_rules_function(__lowerCamelCase , __lowerCamelCase ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , last_epoch=__lowerCamelCase ) def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=-1 )-> List[str]: '''simple docstring''' def lr_lambda(lowerCAmelCase_ ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) return max( 0.0 , float(num_training_steps - current_step ) / float(max(1 , num_training_steps - num_warmup_steps ) ) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 0.5 , lowerCAmelCase_ = -1 )-> List[Any]: '''simple docstring''' def lr_lambda(lowerCAmelCase_ ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) _UpperCAmelCase : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * float(__lowerCamelCase ) * 2.0 * progress )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = 1 , lowerCAmelCase_ = -1 )-> Optional[int]: '''simple docstring''' def lr_lambda(lowerCAmelCase_ ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) _UpperCAmelCase : Dict = float(current_step - num_warmup_steps ) / float(max(1 , num_training_steps - num_warmup_steps ) ) if progress >= 1.0: return 0.0 return max(0.0 , 0.5 * (1.0 + math.cos(math.pi * ((float(__lowerCamelCase ) * progress) % 1.0) )) ) return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1e-7 , lowerCAmelCase_=1.0 , lowerCAmelCase_=-1 )-> List[str]: '''simple docstring''' _UpperCAmelCase : str = optimizer.defaults["""lr"""] if not (lr_init > lr_end): raise ValueError(F'''lr_end ({lr_end}) must be be smaller than initial lr ({lr_init})''' ) def lr_lambda(lowerCAmelCase_ ): if current_step < num_warmup_steps: return float(__lowerCamelCase ) / float(max(1 , __lowerCamelCase ) ) elif current_step > num_training_steps: return lr_end / lr_init # as LambdaLR multiplies by lr_init else: _UpperCAmelCase : Optional[Any] = lr_init - lr_end _UpperCAmelCase : Union[str, Any] = num_training_steps - num_warmup_steps _UpperCAmelCase : Union[str, Any] = 1 - (current_step - num_warmup_steps) / decay_steps _UpperCAmelCase : List[str] = lr_range * pct_remaining**power + lr_end return decay / lr_init # as LambdaLR multiplies by lr_init return LambdaLR(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) A_ : List[Any] = { SchedulerType.LINEAR: get_linear_schedule_with_warmup, SchedulerType.COSINE: get_cosine_schedule_with_warmup, SchedulerType.COSINE_WITH_RESTARTS: get_cosine_with_hard_restarts_schedule_with_warmup, SchedulerType.POLYNOMIAL: get_polynomial_decay_schedule_with_warmup, SchedulerType.CONSTANT: get_constant_schedule, SchedulerType.CONSTANT_WITH_WARMUP: get_constant_schedule_with_warmup, SchedulerType.PIECEWISE_CONSTANT: get_piecewise_constant_schedule, } def snake_case_ ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = None , lowerCAmelCase_ = 1 , lowerCAmelCase_ = 1.0 , lowerCAmelCase_ = -1 , )-> Dict: '''simple docstring''' _UpperCAmelCase : List[Any] = SchedulerType(__lowerCamelCase ) _UpperCAmelCase : List[Any] = TYPE_TO_SCHEDULER_FUNCTION[name] if name == SchedulerType.CONSTANT: return schedule_func(__lowerCamelCase , last_epoch=__lowerCamelCase ) if name == SchedulerType.PIECEWISE_CONSTANT: return schedule_func(__lowerCamelCase , step_rules=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_warmup_steps` if num_warmup_steps is None: raise ValueError(F'''{name} requires `num_warmup_steps`, please provide that argument.''' ) if name == SchedulerType.CONSTANT_WITH_WARMUP: return schedule_func(__lowerCamelCase , num_warmup_steps=__lowerCamelCase , last_epoch=__lowerCamelCase ) # All other schedulers require `num_training_steps` if num_training_steps is None: raise ValueError(F'''{name} requires `num_training_steps`, please provide that argument.''' ) if name == SchedulerType.COSINE_WITH_RESTARTS: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , num_cycles=__lowerCamelCase , last_epoch=__lowerCamelCase , ) if name == SchedulerType.POLYNOMIAL: return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , power=__lowerCamelCase , last_epoch=__lowerCamelCase , ) return schedule_func( __lowerCamelCase , num_warmup_steps=__lowerCamelCase , num_training_steps=__lowerCamelCase , last_epoch=__lowerCamelCase )
215
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def __A ( ) -> Any: a = ArgumentParser( description=( """PyTorch TPU distributed training launch """ """helper utility that will spawn up """ """multiple distributed processes""" ) ) # Optional arguments for the launch helper parser.add_argument("""--num_cores""" , type=__lowerCamelCase , default=1 , help="""Number of TPU cores to use (1 or 8).""" ) # positional parser.add_argument( """training_script""" , type=__lowerCamelCase , help=( """The full path to the single TPU training """ """program/script to be launched in parallel, """ """followed by all the arguments for the """ """training script""" ) , ) # rest from the training program parser.add_argument("""training_script_args""" , nargs=__lowerCamelCase ) return parser.parse_args() def __A ( ) -> Union[str, Any]: a = parse_args() # Import training_script as a module. a = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) a = script_fpath.stem a = importlib.import_module(__lowerCamelCase ) # Patch sys.argv a = [args.training_script] + args.training_script_args + ["""--tpu_num_cores""", str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
228
0
"""simple docstring""" def lowercase_ ( __UpperCAmelCase = "The quick brown fox jumps over the lazy dog" , ) -> bool: lowerCAmelCase__ : Union[str, Any] = set() # Replace all the whitespace in our sentence lowerCAmelCase__ : Any = input_str.replace(""" """ , """""" ) for alpha in input_str: if "a" <= alpha.lower() <= "z": frequency.add(alpha.lower() ) return len(__UpperCAmelCase ) == 26 def lowercase_ ( __UpperCAmelCase = "The quick brown fox jumps over the lazy dog" , ) -> bool: lowerCAmelCase__ : Tuple = [False] * 26 for char in input_str: if char.islower(): lowerCAmelCase__ : int = True elif char.isupper(): lowerCAmelCase__ : int = True return all(__UpperCAmelCase ) def lowercase_ ( __UpperCAmelCase = "The quick brown fox jumps over the lazy dog" , ) -> bool: return len({char for char in input_str.lower() if char.isalpha()} ) == 26 def lowercase_ ( ) -> None: from timeit import timeit lowerCAmelCase__ : List[Any] = """from __main__ import is_pangram, is_pangram_faster, is_pangram_fastest""" print(timeit("""is_pangram()""" , setup=__UpperCAmelCase ) ) print(timeit("""is_pangram_faster()""" , setup=__UpperCAmelCase ) ) print(timeit("""is_pangram_fastest()""" , setup=__UpperCAmelCase ) ) # 5.348480500048026, 2.6477354579837993, 1.8470395830227062 # 5.036091582966037, 2.644472333951853, 1.8869528750656173 if __name__ == "__main__": import doctest doctest.testmod() benchmark()
212
"""simple docstring""" from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def lowercase_ ( __UpperCAmelCase ) -> None: lowerCAmelCase__ , lowerCAmelCase__ : int = analyze_text(__UpperCAmelCase ) lowerCAmelCase__ : Optional[Any] = list(""" """ + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase__ : List[str] = sum(single_char_strings.values() ) # one length string lowerCAmelCase__ : List[str] = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase__ : List[Any] = single_char_strings[ch] lowerCAmelCase__ : int = my_str / all_sum my_fir_sum += prob * math.loga(__UpperCAmelCase ) # entropy formula. # print entropy print(f"""{round(-1 * my_fir_sum ):.1f}""" ) # two len string lowerCAmelCase__ : Tuple = sum(two_char_strings.values() ) lowerCAmelCase__ : str = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase__ : Optional[int] = cha + cha if sequence in two_char_strings: lowerCAmelCase__ : int = two_char_strings[sequence] lowerCAmelCase__ : str = int(__UpperCAmelCase ) / all_sum my_sec_sum += prob * math.loga(__UpperCAmelCase ) # print second entropy print(f"""{round(-1 * my_sec_sum ):.1f}""" ) # print the difference between them print(f"""{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}""" ) def lowercase_ ( __UpperCAmelCase ) -> tuple[dict, dict]: lowerCAmelCase__ : Any = Counter() # type: ignore lowerCAmelCase__ : Tuple = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(__UpperCAmelCase ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def lowercase_ ( ) -> Any: import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
212
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __snake_case ={ """configuration_resnet""": ["""RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ResNetConfig""", """ResNetOnnxConfig"""] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =[ """RESNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """ResNetForImageClassification""", """ResNetModel""", """ResNetPreTrainedModel""", """ResNetBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =[ """TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFResNetForImageClassification""", """TFResNetModel""", """TFResNetPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case =[ """FlaxResNetForImageClassification""", """FlaxResNetModel""", """FlaxResNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys __snake_case =_LazyModule(__name__, globals()["""__file__"""], _import_structure)
4
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __lowerCamelCase ( A__ , A__ , unittest.TestCase ): '''simple docstring''' a_ : Optional[Any] = IFInpaintingSuperResolutionPipeline a_ : int = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""width""", """height"""} a_ : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({"""original_image"""} ) a_ : int = PipelineTesterMixin.required_optional_params - {"""latents"""} def lowerCamelCase ( self : Optional[Any] ): return self._get_superresolution_dummy_components() def lowerCamelCase ( self : Optional[Any] , a_ : List[str] , a_ : Union[str, Any]=0 ): if str(a_ ).startswith("mps" ): lowerCAmelCase_ : List[Any] = torch.manual_seed(a_ ) else: lowerCAmelCase_ : str = torch.Generator(device=a_ ).manual_seed(a_ ) lowerCAmelCase_ : List[str] = floats_tensor((1, 3, 16, 16) , rng=random.Random(a_ ) ).to(a_ ) lowerCAmelCase_ : Tuple = floats_tensor((1, 3, 32, 32) , rng=random.Random(a_ ) ).to(a_ ) lowerCAmelCase_ : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(a_ ) ).to(a_ ) lowerCAmelCase_ : Any = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase ( self : List[Any] ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def lowerCamelCase ( self : Optional[int] ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def lowerCamelCase ( self : Optional[Any] ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def lowerCamelCase ( self : Tuple ): self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def lowerCamelCase ( self : List[str] ): self._test_save_load_local() def lowerCamelCase ( self : Optional[int] ): self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
241
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType lowercase : Tuple = logging.get_logger(__name__) lowercase : str = { 'microsoft/layoutlmv3-base': 'https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json', } class lowerCamelCase__ ( __lowercase): '''simple docstring''' _A = 'layoutlmv3' def __init__( self :List[Any] , a :str=5_0_2_6_5 , a :str=7_6_8 , a :Optional[int]=1_2 , a :str=1_2 , a :Optional[Any]=3_0_7_2 , a :List[Any]="gelu" , a :Any=0.1 , a :Tuple=0.1 , a :Optional[int]=5_1_2 , a :Tuple=2 , a :str=0.02 , a :Any=1E-5 , a :str=1 , a :List[Any]=0 , a :Union[str, Any]=2 , a :Optional[int]=1_0_2_4 , a :Dict=1_2_8 , a :Tuple=1_2_8 , a :Optional[Any]=True , a :Any=3_2 , a :List[Any]=1_2_8 , a :Union[str, Any]=6_4 , a :str=2_5_6 , a :List[str]=True , a :Tuple=True , a :Tuple=True , a :List[str]=2_2_4 , a :Any=3 , a :Dict=1_6 , a :Tuple=None , **a :List[str] , ) -> Tuple: super().__init__( vocab_size=a , hidden_size=a , num_hidden_layers=a , num_attention_heads=a , intermediate_size=a , hidden_act=a , hidden_dropout_prob=a , attention_probs_dropout_prob=a , max_position_embeddings=a , type_vocab_size=a , initializer_range=a , layer_norm_eps=a , pad_token_id=a , bos_token_id=a , eos_token_id=a , **a , ) __UpperCamelCase : List[Any] = max_ad_position_embeddings __UpperCamelCase : Optional[Any] = coordinate_size __UpperCamelCase : List[Any] = shape_size __UpperCamelCase : Optional[Any] = has_relative_attention_bias __UpperCamelCase : Optional[int] = rel_pos_bins __UpperCamelCase : str = max_rel_pos __UpperCamelCase : List[str] = has_spatial_attention_bias __UpperCamelCase : Optional[Any] = rel_ad_pos_bins __UpperCamelCase : Optional[int] = max_rel_ad_pos __UpperCamelCase : List[str] = text_embed __UpperCamelCase : Optional[int] = visual_embed __UpperCamelCase : List[Any] = input_size __UpperCamelCase : Tuple = num_channels __UpperCamelCase : Dict = patch_size __UpperCamelCase : Optional[int] = classifier_dropout class lowerCamelCase__ ( __lowercase): '''simple docstring''' _A = version.parse('1.12') @property def _lowerCamelCase ( self :Tuple ) -> Mapping[str, Mapping[int, str]]: # The order of inputs is different for question answering and sequence classification if self.task in ["question-answering", "sequence-classification"]: return OrderedDict( [ ("input_ids", {0: "batch", 1: "sequence"}), ("attention_mask", {0: "batch", 1: "sequence"}), ("bbox", {0: "batch", 1: "sequence"}), ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) else: return OrderedDict( [ ("input_ids", {0: "batch", 1: "sequence"}), ("bbox", {0: "batch", 1: "sequence"}), ("attention_mask", {0: "batch", 1: "sequence"}), ("pixel_values", {0: "batch", 1: "num_channels"}), ] ) @property def _lowerCamelCase ( self :Dict ) -> float: return 1E-5 @property def _lowerCamelCase ( self :Any ) -> int: return 1_2 def _lowerCamelCase ( self :Any , a :"ProcessorMixin" , a :int = -1 , a :int = -1 , a :bool = False , a :Optional["TensorType"] = None , a :int = 3 , a :int = 4_0 , a :int = 4_0 , ) -> Mapping[str, Any]: setattr(processor.image_processor , "apply_ocr" , a ) # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __UpperCamelCase : Union[str, Any] = compute_effective_axis_dimension( a , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __UpperCamelCase : Optional[int] = processor.tokenizer.num_special_tokens_to_add(a ) __UpperCamelCase : int = compute_effective_axis_dimension( a , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=a ) # Generate dummy inputs according to compute batch and sequence __UpperCamelCase : int = [[" ".join([processor.tokenizer.unk_token] ) * seq_length]] * batch_size # Generate dummy bounding boxes __UpperCamelCase : Union[str, Any] = [[[4_8, 8_4, 7_3, 1_2_8]]] * batch_size # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX # batch_size = compute_effective_axis_dimension(batch_size, fixed_dimension=OnnxConfig.default_fixed_batch) __UpperCamelCase : List[Any] = self._generate_dummy_images(a , a , a , a ) __UpperCamelCase : List[str] = dict( processor( a , text=a , boxes=a , return_tensors=a , ) ) return inputs
151
lowercase : Optional[int] = 9.8_0_6_6_5 def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : float , _lowerCamelCase : float , _lowerCamelCase : float = g) -> float: '''simple docstring''' if fluid_density <= 0: raise ValueError("Impossible fluid density") if volume < 0: raise ValueError("Impossible Object volume") if gravity <= 0: raise ValueError("Impossible Gravity") return fluid_density * gravity * volume if __name__ == "__main__": import doctest # run doctest doctest.testmod()
151
1
import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING __a : Optional[int] = logging.get_logger(__name__) __a : List[str] = { """facebook/detr-resnet-50""": """https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json""", # See all DETR models at https://huggingface.co/models?filter=detr } class _UpperCamelCase ( _UpperCAmelCase ): """simple docstring""" __a : Tuple = '''detr''' __a : Optional[int] = ['''past_key_values'''] __a : Optional[Any] = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self , lowerCAmelCase__=True , lowerCAmelCase__=None , lowerCAmelCase__=3 , lowerCAmelCase__=1_00 , lowerCAmelCase__=6 , lowerCAmelCase__=20_48 , lowerCAmelCase__=8 , lowerCAmelCase__=6 , lowerCAmelCase__=20_48 , lowerCAmelCase__=8 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=True , lowerCAmelCase__="relu" , lowerCAmelCase__=2_56 , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.0 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1.0 , lowerCAmelCase__=False , lowerCAmelCase__="sine" , lowerCAmelCase__="resnet50" , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=1 , lowerCAmelCase__=5 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=1 , lowerCAmelCase__=5 , lowerCAmelCase__=2 , lowerCAmelCase__=0.1 , **lowerCAmelCase__ , ) -> Optional[Any]: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) __lowercase = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): __lowercase = backbone_config.get('''model_type''' ) __lowercase = CONFIG_MAPPING[backbone_model_type] __lowercase = config_class.from_dict(lowerCAmelCase__ ) # set timm attributes to None __lowercase , __lowercase , __lowercase = None, None, None __lowercase = use_timm_backbone __lowercase = backbone_config __lowercase = num_channels __lowercase = num_queries __lowercase = d_model __lowercase = encoder_ffn_dim __lowercase = encoder_layers __lowercase = encoder_attention_heads __lowercase = decoder_ffn_dim __lowercase = decoder_layers __lowercase = decoder_attention_heads __lowercase = dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = activation_function __lowercase = init_std __lowercase = init_xavier_std __lowercase = encoder_layerdrop __lowercase = decoder_layerdrop __lowercase = encoder_layers __lowercase = auxiliary_loss __lowercase = position_embedding_type __lowercase = backbone __lowercase = use_pretrained_backbone __lowercase = dilation # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = mask_loss_coefficient __lowercase = dice_loss_coefficient __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = eos_coefficient super().__init__(is_encoder_decoder=lowerCAmelCase__ , **lowerCAmelCase__ ) @property def _SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return self.encoder_attention_heads @property def _SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return self.d_model @classmethod def _SCREAMING_SNAKE_CASE ( cls , lowerCAmelCase__ , **lowerCAmelCase__ ) -> Optional[int]: '''simple docstring''' return cls(backbone_config=lowerCAmelCase__ , **lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( self ) -> Dict[str, any]: '''simple docstring''' __lowercase = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: __lowercase = self.backbone_config.to_dict() __lowercase = self.__class__.model_type return output class _UpperCamelCase ( _UpperCAmelCase ): """simple docstring""" __a : Union[str, Any] = version.parse('''1.11''' ) @property def _SCREAMING_SNAKE_CASE ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def _SCREAMING_SNAKE_CASE ( self ) -> float: '''simple docstring''' return 1E-5 @property def _SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' return 12
210
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal __a : Optional[Any] = logging.get_logger(__name__) __a : List[str] = TypeVar("""DatasetType""", Dataset, IterableDataset) def UpperCAmelCase ( lowercase , lowercase = None , lowercase = None , lowercase = None , lowercase = None , lowercase = "first_exhausted" , ): """simple docstring""" from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('''Unable to interleave an empty list of datasets.''' ) for i, dataset in enumerate(lowercase ): if not isinstance(lowercase , (Dataset, IterableDataset) ): if isinstance(lowercase , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( F"Dataset at position {i} has at least one split: {list(lowercase )}\n" F"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(lowercase ) )}']" ) raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(lowercase ).__name__}." ) if i == 0: __lowercase , __lowercase = ( (Dataset, IterableDataset) if isinstance(lowercase , lowercase ) else (IterableDataset, Dataset) ) elif not isinstance(lowercase , lowercase ): raise ValueError( F"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(F"{stopping_strategy} is not supported. Please enter a valid stopping_strategy." ) if dataset_type is Dataset: return _interleave_map_style_datasets( lowercase , lowercase , lowercase , info=lowercase , split=lowercase , stopping_strategy=lowercase ) else: return _interleave_iterable_datasets( lowercase , lowercase , lowercase , info=lowercase , split=lowercase , stopping_strategy=lowercase ) def UpperCAmelCase ( lowercase , lowercase = None , lowercase = None , lowercase = 0 , ): """simple docstring""" if not dsets: raise ValueError('''Unable to concatenate an empty list of datasets.''' ) for i, dataset in enumerate(lowercase ): if not isinstance(lowercase , (Dataset, IterableDataset) ): if isinstance(lowercase , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} " '''is an empty dataset dictionary.''' ) raise ValueError( F"Dataset at position {i} has at least one split: {list(lowercase )}\n" F"Please pick one to interleave with the other datasets, for example: dataset['{next(iter(lowercase ) )}']" ) raise ValueError( F"Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(lowercase ).__name__}." ) if i == 0: __lowercase , __lowercase = ( (Dataset, IterableDataset) if isinstance(lowercase , lowercase ) else (IterableDataset, Dataset) ) elif not isinstance(lowercase , lowercase ): raise ValueError( F"Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects." ) if dataset_type is Dataset: return _concatenate_map_style_datasets(lowercase , info=lowercase , split=lowercase , axis=lowercase ) else: return _concatenate_iterable_datasets(lowercase , info=lowercase , split=lowercase , axis=lowercase )
210
1
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __snake_case : def __init__( self : Union[str, Any] , _snake_case : Optional[Any] , _snake_case : Optional[int]=13 , _snake_case : Dict=32 , _snake_case : List[Any]=2 , _snake_case : List[Any]=3 , _snake_case : Tuple=16 , _snake_case : Union[str, Any]=[1, 2, 1] , _snake_case : Optional[Any]=[2, 2, 4] , _snake_case : Any=2 , _snake_case : Optional[Any]=2.0 , _snake_case : List[Any]=True , _snake_case : Any=0.0 , _snake_case : str=0.0 , _snake_case : Optional[Any]=0.1 , _snake_case : int="gelu" , _snake_case : Any=False , _snake_case : str=True , _snake_case : Optional[int]=0.0_2 , _snake_case : str=1e-5 , _snake_case : Tuple=True , _snake_case : Tuple=None , _snake_case : Union[str, Any]=True , _snake_case : str=10 , _snake_case : Union[str, Any]=8 , ): """simple docstring""" UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = image_size UpperCAmelCase_ = patch_size UpperCAmelCase_ = num_channels UpperCAmelCase_ = embed_dim UpperCAmelCase_ = depths UpperCAmelCase_ = num_heads UpperCAmelCase_ = window_size UpperCAmelCase_ = mlp_ratio UpperCAmelCase_ = qkv_bias UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = drop_path_rate UpperCAmelCase_ = hidden_act UpperCAmelCase_ = use_absolute_embeddings UpperCAmelCase_ = patch_norm UpperCAmelCase_ = layer_norm_eps UpperCAmelCase_ = initializer_range UpperCAmelCase_ = is_training UpperCAmelCase_ = scope UpperCAmelCase_ = use_labels UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = encoder_stride def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size) UpperCAmelCase_ = self.get_config() return config, pixel_values, labels def lowerCamelCase ( self : Dict): """simple docstring""" return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def lowerCamelCase ( self : int , _snake_case : int , _snake_case : Dict , _snake_case : Union[str, Any]): """simple docstring""" UpperCAmelCase_ = SwinvaModel(config=_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case) UpperCAmelCase_ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths) - 1)) UpperCAmelCase_ = int(config.embed_dim * 2 ** (len(config.depths) - 1)) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim)) def lowerCamelCase ( self : Any , _snake_case : Union[str, Any] , _snake_case : Any , _snake_case : Tuple): """simple docstring""" UpperCAmelCase_ = SwinvaForMaskedImageModeling(config=_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size)) # test greyscale images UpperCAmelCase_ = 1 UpperCAmelCase_ = SwinvaForMaskedImageModeling(_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) UpperCAmelCase_ = model(_snake_case) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size)) def lowerCamelCase ( self : Any , _snake_case : Tuple , _snake_case : Optional[Any] , _snake_case : str): """simple docstring""" UpperCAmelCase_ = self.type_sequence_label_size UpperCAmelCase_ = SwinvaForImageClassification(_snake_case) model.to(_snake_case) model.eval() UpperCAmelCase_ = model(_snake_case , labels=_snake_case) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) def lowerCamelCase ( self : Any): """simple docstring""" UpperCAmelCase_ = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = config_and_inputs UpperCAmelCase_ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __snake_case ( a , a , unittest.TestCase ): UpperCAmelCase__ : List[str] = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) UpperCAmelCase__ : List[Any] = ( {'''feature-extraction''': SwinvaModel, '''image-classification''': SwinvaForImageClassification} if is_torch_available() else {} ) UpperCAmelCase__ : Dict = False UpperCAmelCase__ : Optional[Any] = False UpperCAmelCase__ : List[Any] = False UpperCAmelCase__ : str = False def lowerCamelCase ( self : List[Any]): """simple docstring""" UpperCAmelCase_ = SwinvaModelTester(self) UpperCAmelCase_ = ConfigTester(self , config_class=_snake_case , embed_dim=37) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case) @unittest.skip(reason='''Got `CUDA error: misaligned address` with PyTorch 2.0.0.''') def lowerCamelCase ( self : Optional[int]): """simple docstring""" pass @unittest.skip(reason='''Swinv2 does not use inputs_embeds''') def lowerCamelCase ( self : int): """simple docstring""" pass def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(_snake_case) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_snake_case , nn.Linear)) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(_snake_case) UpperCAmelCase_ = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _snake_case) def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = True for model_class in self.all_model_classes: UpperCAmelCase_ = True UpperCAmelCase_ = False UpperCAmelCase_ = True UpperCAmelCase_ = model_class(_snake_case) model.to(_snake_case) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(_snake_case , _snake_case)) UpperCAmelCase_ = outputs.attentions UpperCAmelCase_ = len(self.model_tester.depths) self.assertEqual(len(_snake_case) , _snake_case) # check that output_attentions also work using config del inputs_dict["output_attentions"] UpperCAmelCase_ = True UpperCAmelCase_ = config.window_size**2 UpperCAmelCase_ = model_class(_snake_case) model.to(_snake_case) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(_snake_case , _snake_case)) UpperCAmelCase_ = outputs.attentions self.assertEqual(len(_snake_case) , _snake_case) self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) UpperCAmelCase_ = len(_snake_case) # Check attention is always last and order is fine UpperCAmelCase_ = True UpperCAmelCase_ = True UpperCAmelCase_ = model_class(_snake_case) model.to(_snake_case) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(_snake_case , _snake_case)) if hasattr(self.model_tester , '''num_hidden_states_types'''): UpperCAmelCase_ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states UpperCAmelCase_ = 2 self.assertEqual(out_len + added_hidden_states , len(_snake_case)) UpperCAmelCase_ = outputs.attentions self.assertEqual(len(_snake_case) , _snake_case) self.assertListEqual( list(self_attentions[0].shape[-3:]) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def lowerCamelCase ( self : Optional[Any] , _snake_case : Union[str, Any] , _snake_case : Any , _snake_case : Optional[int] , _snake_case : str): """simple docstring""" UpperCAmelCase_ = model_class(_snake_case) model.to(_snake_case) model.eval() with torch.no_grad(): UpperCAmelCase_ = model(**self._prepare_for_class(_snake_case , _snake_case)) UpperCAmelCase_ = outputs.hidden_states UpperCAmelCase_ = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths) + 1) self.assertEqual(len(_snake_case) , _snake_case) # Swinv2 has a different seq_length UpperCAmelCase_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable) else (config.patch_size, config.patch_size) ) UpperCAmelCase_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:]) , [num_patches, self.model_tester.embed_dim] , ) UpperCAmelCase_ = outputs.reshaped_hidden_states self.assertEqual(len(_snake_case) , _snake_case) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = reshaped_hidden_states[0].shape UpperCAmelCase_ = ( reshaped_hidden_states[0].view(_snake_case , _snake_case , height * width).permute(0 , 2 , 1) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:]) , [num_patches, self.model_tester.embed_dim] , ) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: UpperCAmelCase_ = True self.check_hidden_states_output(_snake_case , _snake_case , _snake_case , _snake_case) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ = True self.check_hidden_states_output(_snake_case , _snake_case , _snake_case , _snake_case) def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = 3 UpperCAmelCase_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable) else (self.model_tester.image_size, self.model_tester.image_size) ) UpperCAmelCase_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable) else (config.patch_size, config.patch_size) ) UpperCAmelCase_ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) UpperCAmelCase_ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: UpperCAmelCase_ = True self.check_hidden_states_output(_snake_case , _snake_case , _snake_case , (padded_height, padded_width)) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ = True self.check_hidden_states_output(_snake_case , _snake_case , _snake_case , (padded_height, padded_width)) def lowerCamelCase ( self : int): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_snake_case) def lowerCamelCase ( self : Optional[int]): """simple docstring""" UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_snake_case) @slow def lowerCamelCase ( self : int): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = SwinvaModel.from_pretrained(_snake_case) self.assertIsNotNone(_snake_case) def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase_ = _config_zero_init(_snake_case) for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(config=_snake_case) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class __snake_case ( unittest.TestCase ): @cached_property def lowerCamelCase ( self : List[Any]): """simple docstring""" return ( AutoImageProcessor.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''') if is_vision_available() else None ) @slow def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" UpperCAmelCase_ = SwinvaForImageClassification.from_pretrained('''microsoft/swinv2-tiny-patch4-window8-256''').to( _snake_case) UpperCAmelCase_ = self.default_image_processor UpperCAmelCase_ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''') UpperCAmelCase_ = image_processor(images=_snake_case , return_tensors='''pt''').to(_snake_case) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**_snake_case) # verify the logits UpperCAmelCase_ = torch.Size((1, 1000)) self.assertEqual(outputs.logits.shape , _snake_case) UpperCAmelCase_ = torch.tensor([-0.3_9_4_7, -0.4_3_0_6, 0.0_0_2_6]).to(_snake_case) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _snake_case , atol=1e-4))
7
import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class __snake_case ( a ): UpperCAmelCase__ : Optional[int] = (DPMSolverSinglestepScheduler,) UpperCAmelCase__ : str = (('''num_inference_steps''', 2_5),) def lowerCamelCase ( self : Dict , **_snake_case : Dict): """simple docstring""" UpperCAmelCase_ = { '''num_train_timesteps''': 1000, '''beta_start''': 0.0_0_0_1, '''beta_end''': 0.0_2, '''beta_schedule''': '''linear''', '''solver_order''': 2, '''prediction_type''': '''epsilon''', '''thresholding''': False, '''sample_max_value''': 1.0, '''algorithm_type''': '''dpmsolver++''', '''solver_type''': '''midpoint''', '''lambda_min_clipped''': -float('''inf'''), '''variance_type''': None, } config.update(**_snake_case) return config def lowerCamelCase ( self : Dict , _snake_case : int=0 , **_snake_case : List[Any]): """simple docstring""" UpperCAmelCase_ = dict(self.forward_default_kwargs) UpperCAmelCase_ = kwargs.pop('''num_inference_steps''' , _snake_case) UpperCAmelCase_ = self.dummy_sample UpperCAmelCase_ = 0.1 * sample UpperCAmelCase_ = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ = self.get_scheduler_config(**_snake_case) UpperCAmelCase_ = scheduler_class(**_snake_case) scheduler.set_timesteps(_snake_case) # copy over dummy past residuals UpperCAmelCase_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case) UpperCAmelCase_ = scheduler_class.from_pretrained(_snake_case) new_scheduler.set_timesteps(_snake_case) # copy over dummy past residuals UpperCAmelCase_ = dummy_past_residuals[: new_scheduler.config.solver_order] UpperCAmelCase_ , UpperCAmelCase_ = sample, sample for t in range(_snake_case , time_step + scheduler.config.solver_order + 1): UpperCAmelCase_ = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case).prev_sample UpperCAmelCase_ = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def lowerCamelCase ( self : Tuple): """simple docstring""" pass def lowerCamelCase ( self : Tuple , _snake_case : Optional[Any]=0 , **_snake_case : int): """simple docstring""" UpperCAmelCase_ = dict(self.forward_default_kwargs) UpperCAmelCase_ = kwargs.pop('''num_inference_steps''' , _snake_case) UpperCAmelCase_ = self.dummy_sample UpperCAmelCase_ = 0.1 * sample UpperCAmelCase_ = [residual + 0.2, residual + 0.1_5, residual + 0.1_0] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ = self.get_scheduler_config() UpperCAmelCase_ = scheduler_class(**_snake_case) scheduler.set_timesteps(_snake_case) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase_ = dummy_past_residuals[: scheduler.config.solver_order] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(_snake_case) UpperCAmelCase_ = scheduler_class.from_pretrained(_snake_case) # copy over dummy past residuals new_scheduler.set_timesteps(_snake_case) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase_ = dummy_past_residuals[: new_scheduler.config.solver_order] UpperCAmelCase_ = scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case).prev_sample UpperCAmelCase_ = new_scheduler.step(_snake_case , _snake_case , _snake_case , **_snake_case).prev_sample assert torch.sum(torch.abs(output - new_output)) < 1e-5, "Scheduler outputs are not identical" def lowerCamelCase ( self : Dict , _snake_case : int=None , **_snake_case : Optional[Any]): """simple docstring""" if scheduler is None: UpperCAmelCase_ = self.scheduler_classes[0] UpperCAmelCase_ = self.get_scheduler_config(**_snake_case) UpperCAmelCase_ = scheduler_class(**_snake_case) UpperCAmelCase_ = self.scheduler_classes[0] UpperCAmelCase_ = self.get_scheduler_config(**_snake_case) UpperCAmelCase_ = scheduler_class(**_snake_case) UpperCAmelCase_ = 10 UpperCAmelCase_ = self.dummy_model() UpperCAmelCase_ = self.dummy_sample_deter scheduler.set_timesteps(_snake_case) for i, t in enumerate(scheduler.timesteps): UpperCAmelCase_ = model(_snake_case , _snake_case) UpperCAmelCase_ = scheduler.step(_snake_case , _snake_case , _snake_case).prev_sample return sample def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" UpperCAmelCase_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) UpperCAmelCase_ = 50 UpperCAmelCase_ = self.dummy_model() UpperCAmelCase_ = self.dummy_sample_deter scheduler.set_timesteps(_snake_case) # make sure that the first t is uneven for i, t in enumerate(scheduler.timesteps[3:]): UpperCAmelCase_ = model(_snake_case , _snake_case) UpperCAmelCase_ = scheduler.step(_snake_case , _snake_case , _snake_case).prev_sample UpperCAmelCase_ = torch.mean(torch.abs(_snake_case)) assert abs(result_mean.item() - 0.2_5_7_4) < 1e-3 def lowerCamelCase ( self : int): """simple docstring""" for timesteps in [25, 50, 100, 999, 1000]: self.check_over_configs(num_train_timesteps=_snake_case) def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = DPMSolverSinglestepScheduler(**self.get_scheduler_config()) UpperCAmelCase_ = self.full_loop(scheduler=_snake_case) UpperCAmelCase_ = torch.mean(torch.abs(_snake_case)) assert abs(result_mean.item() - 0.2_7_9_1) < 1e-3 UpperCAmelCase_ = DEISMultistepScheduler.from_config(scheduler.config) UpperCAmelCase_ = DPMSolverMultistepScheduler.from_config(scheduler.config) UpperCAmelCase_ = UniPCMultistepScheduler.from_config(scheduler.config) UpperCAmelCase_ = DPMSolverSinglestepScheduler.from_config(scheduler.config) UpperCAmelCase_ = self.full_loop(scheduler=_snake_case) UpperCAmelCase_ = torch.mean(torch.abs(_snake_case)) assert abs(result_mean.item() - 0.2_7_9_1) < 1e-3 def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" self.check_over_configs(thresholding=_snake_case) for order in [1, 2, 3]: for solver_type in ["midpoint", "heun"]: for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( thresholding=_snake_case , prediction_type=_snake_case , sample_max_value=_snake_case , algorithm_type='''dpmsolver++''' , solver_order=_snake_case , solver_type=_snake_case , ) def lowerCamelCase ( self : Dict): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_snake_case) def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" for algorithm_type in ["dpmsolver", "dpmsolver++"]: for solver_type in ["midpoint", "heun"]: for order in [1, 2, 3]: for prediction_type in ["epsilon", "sample"]: self.check_over_configs( solver_order=_snake_case , solver_type=_snake_case , prediction_type=_snake_case , algorithm_type=_snake_case , ) UpperCAmelCase_ = self.full_loop( solver_order=_snake_case , solver_type=_snake_case , prediction_type=_snake_case , algorithm_type=_snake_case , ) assert not torch.isnan(_snake_case).any(), "Samples have nan numbers" def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" self.check_over_configs(lower_order_final=_snake_case) self.check_over_configs(lower_order_final=_snake_case) def lowerCamelCase ( self : Union[str, Any]): """simple docstring""" self.check_over_configs(lambda_min_clipped=-float('''inf''')) self.check_over_configs(lambda_min_clipped=-5.1) def lowerCamelCase ( self : int): """simple docstring""" self.check_over_configs(variance_type=_snake_case) self.check_over_configs(variance_type='''learned_range''') def lowerCamelCase ( self : Optional[Any]): """simple docstring""" for num_inference_steps in [1, 2, 3, 5, 10, 50, 100, 999, 1000]: self.check_over_forward(num_inference_steps=_snake_case , time_step=0) def lowerCamelCase ( self : str): """simple docstring""" UpperCAmelCase_ = self.full_loop() UpperCAmelCase_ = torch.mean(torch.abs(_snake_case)) assert abs(result_mean.item() - 0.2_7_9_1) < 1e-3 def lowerCamelCase ( self : Optional[Any]): """simple docstring""" UpperCAmelCase_ = self.full_loop(use_karras_sigmas=_snake_case) UpperCAmelCase_ = torch.mean(torch.abs(_snake_case)) assert abs(result_mean.item() - 0.2_2_4_8) < 1e-3 def lowerCamelCase ( self : List[Any]): """simple docstring""" UpperCAmelCase_ = self.full_loop(prediction_type='''v_prediction''') UpperCAmelCase_ = torch.mean(torch.abs(_snake_case)) assert abs(result_mean.item() - 0.1_4_5_3) < 1e-3 def lowerCamelCase ( self : List[str]): """simple docstring""" UpperCAmelCase_ = self.full_loop(prediction_type='''v_prediction''' , use_karras_sigmas=_snake_case) UpperCAmelCase_ = torch.mean(torch.abs(_snake_case)) assert abs(result_mean.item() - 0.0_6_4_9) < 1e-3 def lowerCamelCase ( self : Dict): """simple docstring""" UpperCAmelCase_ = self.scheduler_classes[0] UpperCAmelCase_ = self.get_scheduler_config(thresholding=_snake_case , dynamic_thresholding_ratio=0) UpperCAmelCase_ = scheduler_class(**_snake_case) UpperCAmelCase_ = 10 UpperCAmelCase_ = self.dummy_model() UpperCAmelCase_ = self.dummy_sample_deter.half() scheduler.set_timesteps(_snake_case) for i, t in enumerate(scheduler.timesteps): UpperCAmelCase_ = model(_snake_case , _snake_case) UpperCAmelCase_ = scheduler.step(_snake_case , _snake_case , _snake_case).prev_sample assert sample.dtype == torch.floataa
7
1
'''simple docstring''' import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __lowerCamelCase ( self : Any): '''simple docstring''' __lowercase ='ZinengTang/tvlt-base' __lowercase =tempfile.mkdtemp() def __lowerCamelCase ( self : Dict , **_lowerCAmelCase : List[Any]): '''simple docstring''' return TvltImageProcessor.from_pretrained(self.checkpoint , **A_) def __lowerCamelCase ( self : str , **_lowerCAmelCase : Any): '''simple docstring''' return TvltFeatureExtractor.from_pretrained(self.checkpoint , **A_) def __lowerCamelCase ( self : Optional[Any]): '''simple docstring''' shutil.rmtree(self.tmpdirname) def __lowerCamelCase ( self : List[Any]): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_feature_extractor() __lowercase =TvltProcessor(image_processor=A_ , feature_extractor=A_) processor.save_pretrained(self.tmpdirname) __lowercase =TvltProcessor.from_pretrained(self.tmpdirname) self.assertIsInstance(processor.feature_extractor , A_) self.assertIsInstance(processor.image_processor , A_) def __lowerCamelCase ( self : str): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_feature_extractor() __lowercase =TvltProcessor(image_processor=A_ , feature_extractor=A_) __lowercase =np.ones([1_2_0_0_0]) __lowercase =feature_extractor(A_ , return_tensors='np') __lowercase =processor(audio=A_ , return_tensors='np') for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2) def __lowerCamelCase ( self : int): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_feature_extractor() __lowercase =TvltProcessor(image_processor=A_ , feature_extractor=A_) __lowercase =np.ones([3, 2_2_4, 2_2_4]) __lowercase =image_processor(A_ , return_tensors='np') __lowercase =processor(images=A_ , return_tensors='np') for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2) def __lowerCamelCase ( self : Union[str, Any]): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_feature_extractor() __lowercase =TvltProcessor(image_processor=A_ , feature_extractor=A_) __lowercase =np.ones([1_2_0_0_0]) __lowercase =np.ones([3, 2_2_4, 2_2_4]) __lowercase =processor(audio=A_ , images=A_) self.assertListEqual(list(inputs.keys()) , ['audio_values', 'audio_mask', 'pixel_values', 'pixel_mask']) # test if it raises when no input is passed with pytest.raises(A_): processor() def __lowerCamelCase ( self : Dict): '''simple docstring''' __lowercase =self.get_image_processor() __lowercase =self.get_feature_extractor() __lowercase =TvltProcessor(image_processor=A_ , feature_extractor=A_) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='`processor` and `image_processor`+`feature_extractor` model input names do not match' , )
166
import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home _UpperCAmelCase : Optional[int] = HUGGINGFACE_HUB_CACHE _UpperCAmelCase : List[str] = "config.json" _UpperCAmelCase : Union[str, Any] = "diffusion_pytorch_model.bin" _UpperCAmelCase : List[Any] = "diffusion_flax_model.msgpack" _UpperCAmelCase : Optional[Any] = "model.onnx" _UpperCAmelCase : int = "diffusion_pytorch_model.safetensors" _UpperCAmelCase : Optional[Any] = "weights.pb" _UpperCAmelCase : Tuple = "https://huggingface.co" _UpperCAmelCase : Union[str, Any] = default_cache_path _UpperCAmelCase : Optional[Any] = "diffusers_modules" _UpperCAmelCase : List[Any] = os.getenv("HF_MODULES_CACHE", os.path.join(hf_cache_home, "modules")) _UpperCAmelCase : Tuple = ["fp16", "non-ema"] _UpperCAmelCase : Any = ".self_attn"
222
0
import pprint import requests snake_case_ = 'https://zenquotes.io/api' def lowerCamelCase__ ( ) -> list: return requests.get(API_ENDPOINT_URL + '''/today''' ).json() def lowerCamelCase__ ( ) -> list: return requests.get(API_ENDPOINT_URL + '''/random''' ).json() if __name__ == "__main__": snake_case_ = random_quotes() pprint.pprint(response)
238
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available snake_case_ = { 'configuration_vivit': ['VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'VivitConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = ['VivitImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ 'VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'VivitModel', 'VivitPreTrainedModel', 'VivitForVideoClassification', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys snake_case_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
238
1
def _SCREAMING_SNAKE_CASE ( lowercase : List[str] , lowercase : str ): '''simple docstring''' if a < 0 or b < 0: raise ValueError('the value of both inputs must be positive' ) lowerCamelCase_ = str(bin(_A ) )[2:] # remove the leading "0b" lowerCamelCase_ = str(bin(_A ) )[2:] # remove the leading "0b" lowerCamelCase_ = max(len(_A ) , len(_A ) ) return "0b" + "".join( str(int(char_a != char_b ) ) for char_a, char_b in zip(a_binary.zfill(_A ) , b_binary.zfill(_A ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
204
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging __magic_name__: str = logging.get_logger(__name__) __magic_name__: int = "▁" __magic_name__: List[str] = {"vocab_file": "sentencepiece.bpe.model"} __magic_name__: List[str] = { "vocab_file": { "facebook/nllb-200-distilled-600M": ( "https://huggingface.co/facebook/nllb-200-distilled-600M/blob/main/sentencepiece.bpe.model" ), } } __magic_name__: Tuple = { "facebook/nllb-200-distilled-600M": 1_024, } # fmt: off __magic_name__: int = ["ace_Arab", "ace_Latn", "acm_Arab", "acq_Arab", "aeb_Arab", "afr_Latn", "ajp_Arab", "aka_Latn", "amh_Ethi", "apc_Arab", "arb_Arab", "ars_Arab", "ary_Arab", "arz_Arab", "asm_Beng", "ast_Latn", "awa_Deva", "ayr_Latn", "azb_Arab", "azj_Latn", "bak_Cyrl", "bam_Latn", "ban_Latn", "bel_Cyrl", "bem_Latn", "ben_Beng", "bho_Deva", "bjn_Arab", "bjn_Latn", "bod_Tibt", "bos_Latn", "bug_Latn", "bul_Cyrl", "cat_Latn", "ceb_Latn", "ces_Latn", "cjk_Latn", "ckb_Arab", "crh_Latn", "cym_Latn", "dan_Latn", "deu_Latn", "dik_Latn", "dyu_Latn", "dzo_Tibt", "ell_Grek", "eng_Latn", "epo_Latn", "est_Latn", "eus_Latn", "ewe_Latn", "fao_Latn", "pes_Arab", "fij_Latn", "fin_Latn", "fon_Latn", "fra_Latn", "fur_Latn", "fuv_Latn", "gla_Latn", "gle_Latn", "glg_Latn", "grn_Latn", "guj_Gujr", "hat_Latn", "hau_Latn", "heb_Hebr", "hin_Deva", "hne_Deva", "hrv_Latn", "hun_Latn", "hye_Armn", "ibo_Latn", "ilo_Latn", "ind_Latn", "isl_Latn", "ita_Latn", "jav_Latn", "jpn_Jpan", "kab_Latn", "kac_Latn", "kam_Latn", "kan_Knda", "kas_Arab", "kas_Deva", "kat_Geor", "knc_Arab", "knc_Latn", "kaz_Cyrl", "kbp_Latn", "kea_Latn", "khm_Khmr", "kik_Latn", "kin_Latn", "kir_Cyrl", "kmb_Latn", "kon_Latn", "kor_Hang", "kmr_Latn", "lao_Laoo", "lvs_Latn", "lij_Latn", "lim_Latn", "lin_Latn", "lit_Latn", "lmo_Latn", "ltg_Latn", "ltz_Latn", "lua_Latn", "lug_Latn", "luo_Latn", "lus_Latn", "mag_Deva", "mai_Deva", "mal_Mlym", "mar_Deva", "min_Latn", "mkd_Cyrl", "plt_Latn", "mlt_Latn", "mni_Beng", "khk_Cyrl", "mos_Latn", "mri_Latn", "zsm_Latn", "mya_Mymr", "nld_Latn", "nno_Latn", "nob_Latn", "npi_Deva", "nso_Latn", "nus_Latn", "nya_Latn", "oci_Latn", "gaz_Latn", "ory_Orya", "pag_Latn", "pan_Guru", "pap_Latn", "pol_Latn", "por_Latn", "prs_Arab", "pbt_Arab", "quy_Latn", "ron_Latn", "run_Latn", "rus_Cyrl", "sag_Latn", "san_Deva", "sat_Beng", "scn_Latn", "shn_Mymr", "sin_Sinh", "slk_Latn", "slv_Latn", "smo_Latn", "sna_Latn", "snd_Arab", "som_Latn", "sot_Latn", "spa_Latn", "als_Latn", "srd_Latn", "srp_Cyrl", "ssw_Latn", "sun_Latn", "swe_Latn", "swh_Latn", "szl_Latn", "tam_Taml", "tat_Cyrl", "tel_Telu", "tgk_Cyrl", "tgl_Latn", "tha_Thai", "tir_Ethi", "taq_Latn", "taq_Tfng", "tpi_Latn", "tsn_Latn", "tso_Latn", "tuk_Latn", "tum_Latn", "tur_Latn", "twi_Latn", "tzm_Tfng", "uig_Arab", "ukr_Cyrl", "umb_Latn", "urd_Arab", "uzn_Latn", "vec_Latn", "vie_Latn", "war_Latn", "wol_Latn", "xho_Latn", "ydd_Hebr", "yor_Latn", "yue_Hant", "zho_Hans", "zho_Hant", "zul_Latn"] class snake_case__ ( _lowerCAmelCase ): lowercase__ : str = VOCAB_FILES_NAMES lowercase__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : str = ['''input_ids''', '''attention_mask'''] lowercase__ : List[int] = [] lowercase__ : List[int] = [] def __init__( self , lowerCAmelCase__ , lowerCAmelCase__="<s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="</s>" , lowerCAmelCase__="<s>" , lowerCAmelCase__="<unk>" , lowerCAmelCase__="<pad>" , lowerCAmelCase__="<mask>" , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__ = None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> int: # Mask token behave like a normal word, i.e. include the space before it __magic_name__ : Any = AddedToken(lowerCAmelCase__ , lstrip=lowerCAmelCase__ , rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else mask_token __magic_name__ : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs __magic_name__ : Optional[Any] = legacy_behaviour super().__init__( bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , src_lang=lowerCAmelCase__ , tgt_lang=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , sp_model_kwargs=self.sp_model_kwargs , legacy_behaviour=lowerCAmelCase__ , **lowerCAmelCase__ , ) __magic_name__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCAmelCase__ ) ) __magic_name__ : List[Any] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | ---- | ---- | ---- | ---- | ---- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' # spm | '<unk>' | '<s>' | '</s>' | 'an' | '▁n' | '▁m' | '▁t' | '▁k' | '▁a' | '▁s' # Mimic fairseq token-to-id alignment for the first 4 token __magic_name__ : List[str] = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __magic_name__ : List[Any] = 1 __magic_name__ : Dict = len(self.sp_model ) __magic_name__ : int = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(lowerCAmelCase__ ) } __magic_name__ : Optional[int] = {v: k for k, v in self.lang_code_to_id.items()} __magic_name__ : Union[str, Any] = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) __magic_name__ : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} __magic_name__ : List[str] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) __magic_name__ : List[Any] = src_lang if src_lang is not None else """eng_Latn""" __magic_name__ : Any = self.lang_code_to_id[self._src_lang] __magic_name__ : Optional[int] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ) -> Any: __magic_name__ : List[Any] = self.__dict__.copy() __magic_name__ : int = None __magic_name__ : Optional[int] = self.sp_model.serialized_model_proto() return state def __setstate__( self , lowerCAmelCase__ ) -> Any: __magic_name__ : Any = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __magic_name__ : Any = {} __magic_name__ : Union[str, Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def __magic_name__ ( self ) -> str: return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def __magic_name__ ( self ) -> str: return self._src_lang @src_lang.setter def __magic_name__ ( self , lowerCAmelCase__ ) -> None: __magic_name__ : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None , lowerCAmelCase__ = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__ ) __magic_name__ : Optional[int] = [1] * len(self.prefix_tokens ) __magic_name__ : Any = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase__ )) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase__ )) + ([0] * len(lowerCAmelCase__ )) + suffix_ones def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: __magic_name__ : str = [self.sep_token_id] __magic_name__ : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) -> List[str]: if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) __magic_name__ : Dict = src_lang __magic_name__ : List[Any] = self(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) __magic_name__ : Optional[Any] = self.convert_tokens_to_ids(lowerCAmelCase__ ) __magic_name__ : Tuple = tgt_lang_id return inputs def __magic_name__ ( self ) -> int: __magic_name__ : str = {self.convert_ids_to_tokens(lowerCAmelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __magic_name__ ( self , lowerCAmelCase__ ) -> List[str]: return self.sp_model.encode(lowerCAmelCase__ , out_type=lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Optional[Any]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __magic_name__ : List[str] = self.sp_model.PieceToId(lowerCAmelCase__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __magic_name__ ( self , lowerCAmelCase__ ) -> str: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __magic_name__ ( self , lowerCAmelCase__ ) -> Union[str, Any]: __magic_name__ : Tuple = """""".join(lowerCAmelCase__ ).replace(lowerCAmelCase__ , """ """ ).strip() return out_string def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase__ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __magic_name__ : List[Any] = os.path.join( lowerCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase__ , """wb""" ) as fi: __magic_name__ : List[str] = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase__ ) return (out_vocab_file,) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = "eng_Latn" , lowerCAmelCase__ = None , lowerCAmelCase__ = "fra_Latn" , **lowerCAmelCase__ , ) -> BatchEncoding: __magic_name__ : List[str] = src_lang __magic_name__ : Dict = tgt_lang return super().prepare_seqaseq_batch(lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ ) def __magic_name__ ( self ) -> str: return self.set_src_lang_special_tokens(self.src_lang ) def __magic_name__ ( self ) -> List[str]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __magic_name__ ( self , lowerCAmelCase__ ) -> None: __magic_name__ : Optional[int] = self.lang_code_to_id[src_lang] if self.legacy_behaviour: __magic_name__ : List[str] = [] __magic_name__ : Tuple = [self.eos_token_id, self.cur_lang_code] else: __magic_name__ : str = [self.cur_lang_code] __magic_name__ : List[Any] = [self.eos_token_id] def __magic_name__ ( self , lowerCAmelCase__ ) -> None: __magic_name__ : List[str] = self.lang_code_to_id[lang] if self.legacy_behaviour: __magic_name__ : List[str] = [] __magic_name__ : Optional[int] = [self.eos_token_id, self.cur_lang_code] else: __magic_name__ : Optional[int] = [self.cur_lang_code] __magic_name__ : Union[str, Any] = [self.eos_token_id]
342
0
import os import posixpath import uuid from dataclasses import dataclass from typing import TYPE_CHECKING, Iterable, List, Optional, Tuple, Union import numpy as np import pyarrow as pa import datasets from datasets.arrow_writer import ArrowWriter, ParquetWriter from datasets.config import MAX_SHARD_SIZE from datasets.filesystems import ( is_remote_filesystem, rename, ) from datasets.iterable_dataset import _BaseExamplesIterable from datasets.utils.py_utils import convert_file_size_to_int lowerCAmelCase = datasets.utils.logging.get_logger(__name__) if TYPE_CHECKING: import pyspark @dataclass class A ( datasets.BuilderConfig ): UpperCamelCase_ : Optional[datasets.Features] =None def _lowerCamelCase( lowercase__ , lowercase__ , ) -> List[str]: '''simple docstring''' import pyspark def generate_fn(): __lowercase= df.select('*' , pyspark.sql.functions.spark_partition_id().alias('part_id' ) ) for partition_id in partition_order: __lowercase= df_with_partition_id.select('*' ).where(F'part_id = {partition_id}' ).drop('part_id' ) __lowercase= partition_df.collect() __lowercase= 0 for row in rows: yield F'{partition_id}_{row_id}', row.asDict() row_id += 1 return generate_fn class A ( _BaseExamplesIterable ): def __init__(self , lowerCAmelCase , lowerCAmelCase=None , ): __lowercase= df __lowercase= partition_order or range(self.df.rdd.getNumPartitions() ) __lowercase= _generate_iterable_examples(self.df , self.partition_order ) def __iter__(self ): yield from self.generate_examples_fn() def _A (self , lowerCAmelCase ): __lowercase= list(range(self.df.rdd.getNumPartitions() ) ) generator.shuffle(lowerCAmelCase ) return SparkExamplesIterable(self.df , partition_order=lowerCAmelCase ) def _A (self , lowerCAmelCase , lowerCAmelCase ): __lowercase= self.split_shard_indices_by_worker(lowerCAmelCase , lowerCAmelCase ) return SparkExamplesIterable(self.df , partition_order=lowerCAmelCase ) @property def _A (self ): return len(self.partition_order ) class A ( datasets.DatasetBuilder ): UpperCamelCase_ : Optional[Any] =SparkConfig def __init__(self , lowerCAmelCase , lowerCAmelCase = None , lowerCAmelCase = None , **lowerCAmelCase , ): import pyspark __lowercase= pyspark.sql.SparkSession.builder.getOrCreate() __lowercase= df __lowercase= working_dir super().__init__( cache_dir=lowerCAmelCase , config_name=str(self.df.semanticHash() ) , **lowerCAmelCase , ) def _A (self ): # Returns the path of the created file. def create_cache_and_write_probe(lowerCAmelCase ): # makedirs with exist_ok will recursively create the directory. It will not throw an error if directories # already exist. os.makedirs(self._cache_dir , exist_ok=lowerCAmelCase ) __lowercase= os.path.join(self._cache_dir , 'fs_test' + uuid.uuida().hex ) # Opening the file in append mode will create a new file unless it already exists, in which case it will not # change the file contents. open(lowerCAmelCase , 'a' ) return [probe_file] if self._spark.conf.get('spark.master' , '' ).startswith('local' ): return # If the cluster is multi-node, make sure that the user provided a cache_dir and that it is on an NFS # accessible to the driver. # TODO: Stream batches to the driver using ArrowCollectSerializer instead of throwing an error. if self._cache_dir: __lowercase= ( self._spark.sparkContext.parallelize(range(1 ) , 1 ).mapPartitions(lowerCAmelCase ).collect() ) if os.path.isfile(probe[0] ): return raise ValueError( 'When using Dataset.from_spark on a multi-node cluster, the driver and all workers should be able to access cache_dir' ) def _A (self ): return datasets.DatasetInfo(features=self.config.features ) def _A (self , lowerCAmelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN )] def _A (self , lowerCAmelCase ): import pyspark def get_arrow_batch_size(lowerCAmelCase ): for batch in it: yield pa.RecordBatch.from_pydict({'batch_bytes': [batch.nbytes]} ) __lowercase= self.df.count() __lowercase= df_num_rows if df_num_rows <= 1_0_0 else 1_0_0 # Approximate the size of each row (in Arrow format) by averaging over a max-100-row sample. __lowercase= ( self.df.limit(lowerCAmelCase ) .repartition(1 ) .mapInArrow(lowerCAmelCase , 'batch_bytes: long' ) .agg(pyspark.sql.functions.sum('batch_bytes' ).alias('sample_bytes' ) ) .collect()[0] .sample_bytes / sample_num_rows ) __lowercase= approx_bytes_per_row * df_num_rows if approx_total_size > max_shard_size: # Make sure there is at least one row per partition. __lowercase= min(lowerCAmelCase , int(approx_total_size / max_shard_size ) ) __lowercase= self.df.repartition(lowerCAmelCase ) def _A (self , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ): import pyspark __lowercase= ParquetWriter if file_format == 'parquet' else ArrowWriter __lowercase= os.path.join(self._working_dir , os.path.basename(lowerCAmelCase ) ) if self._working_dir else fpath __lowercase= file_format == 'parquet' # Define these so that we don't reference self in write_arrow, which will result in a pickling error due to # pickling the SparkContext. __lowercase= self.config.features __lowercase= self._writer_batch_size __lowercase= self._fs.storage_options def write_arrow(lowerCAmelCase ): # Within the same SparkContext, no two task attempts will share the same attempt ID. __lowercase= pyspark.TaskContext().taskAttemptId() __lowercase= next(lowerCAmelCase , lowerCAmelCase ) if first_batch is None: # Some partitions might not receive any data. return pa.RecordBatch.from_arrays( [[task_id], [0], [0]] , names=['task_id', 'num_examples', 'num_bytes'] , ) __lowercase= 0 __lowercase= writer_class( features=lowerCAmelCase , path=working_fpath.replace('SSSSS' , f'{shard_id:05d}' ).replace('TTTTT' , f'{task_id:05d}' ) , writer_batch_size=lowerCAmelCase , storage_options=lowerCAmelCase , embed_local_files=lowerCAmelCase , ) __lowercase= pa.Table.from_batches([first_batch] ) writer.write_table(lowerCAmelCase ) for batch in it: if max_shard_size is not None and writer._num_bytes >= max_shard_size: __lowercase, __lowercase= writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]] , names=['task_id', 'num_examples', 'num_bytes'] , ) shard_id += 1 __lowercase= writer_class( features=writer._features , path=working_fpath.replace('SSSSS' , f'{shard_id:05d}' ).replace('TTTTT' , f'{task_id:05d}' ) , writer_batch_size=lowerCAmelCase , storage_options=lowerCAmelCase , embed_local_files=lowerCAmelCase , ) __lowercase= pa.Table.from_batches([batch] ) writer.write_table(lowerCAmelCase ) if writer._num_bytes > 0: __lowercase, __lowercase= writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]] , names=['task_id', 'num_examples', 'num_bytes'] , ) if working_fpath != fpath: for file in os.listdir(os.path.dirname(lowerCAmelCase ) ): __lowercase= os.path.join(os.path.dirname(lowerCAmelCase ) , os.path.basename(lowerCAmelCase ) ) shutil.move(lowerCAmelCase , lowerCAmelCase ) __lowercase= ( self.df.mapInArrow(lowerCAmelCase , 'task_id: long, num_examples: long, num_bytes: long' ) .groupBy('task_id' ) .agg( pyspark.sql.functions.sum('num_examples' ).alias('total_num_examples' ) , pyspark.sql.functions.sum('num_bytes' ).alias('total_num_bytes' ) , pyspark.sql.functions.count('num_bytes' ).alias('num_shards' ) , pyspark.sql.functions.collect_list('num_examples' ).alias('shard_lengths' ) , ) .collect() ) for row in stats: yield row.task_id, (row.total_num_examples, row.total_num_bytes, row.num_shards, row.shard_lengths) def _A (self , lowerCAmelCase , lowerCAmelCase = "arrow" , lowerCAmelCase = None , lowerCAmelCase = None , **lowerCAmelCase , ): self._validate_cache_dir() __lowercase= convert_file_size_to_int(max_shard_size or MAX_SHARD_SIZE ) self._repartition_df_if_needed(lowerCAmelCase ) __lowercase= not is_remote_filesystem(self._fs ) __lowercase= os.path.join if is_local else posixpath.join __lowercase= '-TTTTT-SSSSS-of-NNNNN' __lowercase= f'{self.name}-{split_generator.name}{SUFFIX}.{file_format}' __lowercase= path_join(self._output_dir , lowerCAmelCase ) __lowercase= 0 __lowercase= 0 __lowercase= 0 __lowercase= [] __lowercase= [] for task_id, content in self._prepare_split_single(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ): ( ( __lowercase ), ( __lowercase ), ( __lowercase ), ( __lowercase ), )= content if num_bytes > 0: total_num_examples += num_examples total_num_bytes += num_bytes total_shards += num_shards task_id_and_num_shards.append((task_id, num_shards) ) all_shard_lengths.extend(lowerCAmelCase ) __lowercase= total_num_examples __lowercase= total_num_bytes # should rename everything at the end logger.debug(f'Renaming {total_shards} shards.' ) if total_shards > 1: __lowercase= all_shard_lengths # Define fs outside of _rename_shard so that we don't reference self in the function, which will result in a # pickling error due to pickling the SparkContext. __lowercase= self._fs # use the -SSSSS-of-NNNNN pattern def _rename_shard( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , ): rename( lowerCAmelCase , fpath.replace('SSSSS' , f'{shard_id:05d}' ).replace('TTTTT' , f'{task_id:05d}' ) , fpath.replace('TTTTT-SSSSS' , f'{global_shard_id:05d}' ).replace('NNNNN' , f'{total_shards:05d}' ) , ) __lowercase= [] __lowercase= 0 for i in range(len(lowerCAmelCase ) ): __lowercase, __lowercase= task_id_and_num_shards[i] for shard_id in range(lowerCAmelCase ): args.append([task_id, shard_id, global_shard_id] ) global_shard_id += 1 self._spark.sparkContext.parallelize(lowerCAmelCase , len(lowerCAmelCase ) ).map(lambda lowerCAmelCase : _rename_shard(*lowerCAmelCase ) ).collect() else: # don't use any pattern __lowercase= 0 __lowercase= task_id_and_num_shards[0][0] self._rename( fpath.replace('SSSSS' , f'{shard_id:05d}' ).replace('TTTTT' , f'{task_id:05d}' ) , fpath.replace(lowerCAmelCase , '' ) , ) def _A (self , lowerCAmelCase , ): return SparkExamplesIterable(self.df )
352
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def _lowerCamelCase( lowercase__ , lowercase__ ) -> Optional[int]: '''simple docstring''' __lowercase= 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg' __lowercase= Image.open(requests.get(lowercase__ , stream=lowercase__ ).raw ).convert('RGB' ) __lowercase= transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.4814_5466, 0.457_8275, 0.4082_1073) , (0.2686_2954, 0.2613_0258, 0.2757_7711) ), ] ) __lowercase= transform(lowercase__ ).unsqueeze(0 ).to(lowercase__ ) return image def _lowerCamelCase( lowercase__ ) -> Dict: '''simple docstring''' if "visual_encoder" in key: __lowercase= re.sub('visual_encoder*' , 'vision_model.encoder' , lowercase__ ) if "blocks" in key: __lowercase= re.sub(R'blocks' , 'layers' , lowercase__ ) if "attn" in key: __lowercase= re.sub(R'attn' , 'self_attn' , lowercase__ ) if "norm1" in key: __lowercase= re.sub(R'norm1' , 'layer_norm1' , lowercase__ ) if "norm2" in key: __lowercase= re.sub(R'norm2' , 'layer_norm2' , lowercase__ ) if "encoder.norm" in key: __lowercase= re.sub(R'encoder.norm' , 'post_layernorm' , lowercase__ ) if "encoder.patch_embed.proj" in key: __lowercase= re.sub(R'encoder.patch_embed.proj' , 'embeddings.patch_embedding' , lowercase__ ) if "encoder.pos_embed" in key: __lowercase= re.sub(R'encoder.pos_embed' , 'embeddings.position_embedding' , lowercase__ ) if "encoder.cls_token" in key: __lowercase= re.sub(R'encoder.cls_token' , 'embeddings.class_embedding' , lowercase__ ) if "self_attn" in key: __lowercase= re.sub(R'self_attn.proj' , 'self_attn.projection' , lowercase__ ) return key @torch.no_grad() def _lowerCamelCase( lowercase__ , lowercase__=None ) -> int: '''simple docstring''' if config_path is not None: __lowercase= BlipConfig.from_pretrained(lowercase__ ) else: __lowercase= BlipConfig(projection_dim=5_1_2 , text_config={} , vision_config={} ) __lowercase= BlipForConditionalGeneration(lowercase__ ).eval() __lowercase= 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth' __lowercase= blip_decoder(pretrained=lowercase__ , image_size=3_8_4 , vit='base' ) __lowercase= pt_model.eval() __lowercase= pt_model.state_dict() for key in modified_state_dict.copy(): __lowercase= modified_state_dict.pop(lowercase__ ) __lowercase= rename_key(lowercase__ ) __lowercase= value hf_model.load_state_dict(lowercase__ ) __lowercase= 3_8_4 __lowercase= load_demo_image(image_size=lowercase__ , device='cpu' ) __lowercase= BertTokenizer.from_pretrained('bert-base-uncased' ) __lowercase= tokenizer(['a picture of'] ).input_ids __lowercase= hf_model.generate(lowercase__ , lowercase__ ) assert out[0].tolist() == [3_0_5_2_2, 1_0_3_7, 3_8_6_1, 1_9_9_7, 1_0_3_7, 2_4_5_0, 3_5_6_4, 2_0_0_6, 1_9_9_6, 3_5_0_9, 2_0_0_7, 2_0_1_4, 3_8_9_9, 1_0_2] __lowercase= hf_model.generate(lowercase__ ) assert out[0].tolist() == [3_0_5_2_2, 1_0_3_7, 2_4_5_0, 3_5_6_4, 2_0_0_6, 1_9_9_6, 3_5_0_9, 2_0_0_7, 2_0_1_4, 3_8_9_9, 1_0_2] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(lowercase__ ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' __lowercase= ( 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth' ) __lowercase= blip_vqa(pretrained=lowercase__ , image_size=lowercase__ , vit='base' ) vqa_model.eval() __lowercase= vqa_model.state_dict() for key in modified_state_dict.copy(): __lowercase= modified_state_dict.pop(lowercase__ ) __lowercase= rename_key(lowercase__ ) __lowercase= value __lowercase= BlipForQuestionAnswering(lowercase__ ) hf_vqa_model.load_state_dict(lowercase__ ) __lowercase= ['How many dogs are in this image?'] __lowercase= tokenizer(lowercase__ , return_tensors='pt' ).input_ids __lowercase= hf_vqa_model.generate(lowercase__ , lowercase__ ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + '_vqa' ) __lowercase= 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth' __lowercase= blip_itm(pretrained=lowercase__ , image_size=lowercase__ , vit='base' ) itm_model.eval() __lowercase= itm_model.state_dict() for key in modified_state_dict.copy(): __lowercase= modified_state_dict.pop(lowercase__ ) __lowercase= rename_key(lowercase__ ) __lowercase= value __lowercase= BlipForImageTextRetrieval(lowercase__ ) __lowercase= ['A picture of a woman with a dog sitting in a beach'] __lowercase= tokenizer( lowercase__ , return_tensors='pt' , padding='max_length' , truncation=lowercase__ , max_length=3_5 , ).input_ids hf_itm_model.load_state_dict(lowercase__ ) hf_itm_model.eval() __lowercase= hf_itm_model(lowercase__ , lowercase__ , use_itm_head=lowercase__ ) __lowercase= hf_itm_model(lowercase__ , lowercase__ , use_itm_head=lowercase__ ) assert out[0].item() == 0.2110_6874_9427_7954 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.4_5698_8453_8650_5127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + '_itm' ) if __name__ == "__main__": lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') lowerCAmelCase = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
304
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase: int = logging.get_logger(__name__) lowerCAmelCase: str = torch.device('cpu') def lowerCamelCase__ ( ): a : str = 'http://images.cocodataset.org/val2017/000000039769.jpg' a : List[Any] = Image.open(requests.get(_A , stream=_A ).raw ) return im def lowerCamelCase__ ( _A ): if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1703E00, 2.1107E00, -2.0811E00, 8.8685E-01, 2.4360E-01] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9636E-01, 2.3478E-01, -1.6963E00, -1.7381E00, -8.6337E-01] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2768E-01, -4.7429E-01, -1.0897E00, -1.0248E00, 3.5523E-02] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5330E-01, 2.4211E-01, -6.0185E-01, -8.2789E-01, -6.0446E-02] ) def lowerCamelCase__ ( _A , _A , _A ): a : Any = dct.pop(_A ) a : List[Any] = val def lowerCamelCase__ ( _A ): a : List[str] = [] for k in state_dict.keys(): a : List[str] = k if ".pwconv" in k: a : Tuple = k_new.replace('.pwconv' , '.point_wise_conv' ) if ".dwconv" in k: a : str = k_new.replace('.dwconv' , '.depth_wise_conv' ) if ".Proj." in k: a : Optional[Any] = k_new.replace('.Proj.' , '.proj.' ) if "patch_embed" in k_new: a : int = k_new.replace('patch_embed' , 'swiftformer.patch_embed.patch_embedding' ) if "network" in k_new: a : Optional[Any] = k_new.split('.' ) if ls[2].isdigit(): a : Any = 'swiftformer.encoder.network.' + ls[1] + '.blocks.' + ls[2] + '.' + '.'.join(ls[3:] ) else: a : Optional[int] = k_new.replace('network' , 'swiftformer.encoder.network' ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowerCamelCase__ ( _A , _A , _A ): a : Any = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size a : Optional[Any] = 1000 a : Optional[Any] = 'huggingface/label-files' a : List[Any] = 'imagenet-1k-id2label.json' a : int = json.load(open(hf_hub_download(_A , _A , repo_type='dataset' ) , 'r' ) ) a : Dict = {int(_A ): v for k, v in idalabel.items()} a : Union[str, Any] = idalabel a : str = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": a : List[str] = [3, 3, 6, 4] a : Optional[int] = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": a : Dict = [3, 3, 9, 6] a : List[Any] = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": a : Any = [4, 3, 10, 5] a : Tuple = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": a : Optional[int] = [4, 4, 12, 6] a : Tuple = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith('https' ): a : Tuple = torch.hub.load_state_dict_from_url(_A , map_location='cpu' , check_hash=_A ) else: a : Any = torch.load(_A , map_location='cpu' ) a : Union[str, Any] = checkpoint a : Optional[Any] = create_rename_keys(_A ) for rename_key_src, rename_key_dest in rename_keys: rename_key(_A , _A , _A ) # load HuggingFace model a : Optional[int] = SwiftFormerForImageClassification(_A ).eval() hf_model.load_state_dict(_A ) # prepare test inputs a : Dict = prepare_img() a : int = ViTImageProcessor.from_pretrained('preprocessor_config' ) a : List[str] = processor(images=_A , return_tensors='pt' ) # compare outputs from both models a : str = get_expected_output(_A ) a : Optional[Any] = hf_model(inputs['pixel_values'] ).logits assert hf_logits.shape == torch.Size([1, 1000] ) assert torch.allclose(hf_logits[0, 0:5] , _A , atol=1E-3 ) Path(_A ).mkdir(exist_ok=_A ) print(f"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(_A ) if __name__ == "__main__": lowerCAmelCase: Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') lowerCAmelCase: Tuple = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
297
'''simple docstring''' import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) lowerCAmelCase: int = [ 'cross_validation.py', 'gradient_accumulation.py', 'local_sgd.py', 'multi_process_metrics.py', 'memory.py', 'automatic_gradient_accumulation.py', 'fsdp_with_peak_mem_tracking.py', 'deepspeed_with_config_support.py', 'megatron_lm_gpt_pretraining.py', ] class a__( unittest.TestCase ): def lowercase_ ( self : int , __snake_case : str , __snake_case : bool , __snake_case : str = None , __snake_case : list = None ): a : Optional[int] = None a : Tuple = os.path.abspath(os.path.join('examples' , 'by_feature' ) ) a : List[str] = os.path.abspath('examples' ) for item in os.listdir(__snake_case ): if item not in EXCLUDE_EXAMPLES: a : int = os.path.join(__snake_case , __snake_case ) if os.path.isfile(__snake_case ) and ".py" in item_path: with self.subTest( tested_script=__snake_case , feature_script=__snake_case , tested_section='main()' if parser_only else 'training_function()' , ): a : List[Any] = compare_against_test( os.path.join(__snake_case , __snake_case ) , __snake_case , __snake_case , __snake_case ) a : Union[str, Any] = '\n'.join(__snake_case ) if special_strings is not None: for string in special_strings: a : Union[str, Any] = diff.replace(__snake_case , '' ) self.assertEqual(__snake_case , '' ) def lowercase_ ( self : Optional[Any] ): self.one_complete_example('complete_nlp_example.py' , __snake_case ) self.one_complete_example('complete_nlp_example.py' , __snake_case ) def lowercase_ ( self : Any ): a : Dict = os.path.abspath(os.path.join('examples' , 'cv_example.py' ) ) a : int = [ ' ' * 16 + '{\n\n', ' ' * 20 + '"accuracy": eval_metric["accuracy"],\n\n', ' ' * 20 + '"f1": eval_metric["f1"],\n\n', ' ' * 20 + '"train_loss": total_loss.item() / len(train_dataloader),\n\n', ' ' * 20 + '"epoch": epoch,\n\n', ' ' * 16 + '},\n\n', ' ' * 16 + 'step=epoch,\n', ' ' * 12, ' ' * 8 + 'for step, batch in enumerate(active_dataloader):\n', ] self.one_complete_example('complete_cv_example.py' , __snake_case , __snake_case , __snake_case ) self.one_complete_example('complete_cv_example.py' , __snake_case , __snake_case , __snake_case ) @mock.patch.dict(os.environ , {"""TESTING_MOCKED_DATALOADERS""": """1"""} ) class a__( lowerCamelCase__ ): lowercase__ = False @classmethod def lowercase_ ( cls : Optional[int] ): super().setUpClass() a : List[str] = tempfile.mkdtemp() a : Tuple = os.path.join(cls._tmpdir , 'default_config.yml' ) write_basic_config(save_location=cls.configPath ) a : Optional[int] = ['accelerate', 'launch', '--config_file', cls.configPath] @classmethod def lowercase_ ( cls : Optional[int] ): super().tearDownClass() shutil.rmtree(cls._tmpdir ) def lowercase_ ( self : Tuple ): a : Union[str, Any] = F""" examples/by_feature/checkpointing.py --checkpointing_steps epoch --output_dir {self.tmpdir} """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'epoch_0' ) ) ) def lowercase_ ( self : Dict ): a : Union[str, Any] = F""" examples/by_feature/checkpointing.py --checkpointing_steps 1 --output_dir {self.tmpdir} """.split() a : int = run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'step_2' ) ) ) def lowercase_ ( self : Any ): a : Tuple = F""" examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , "epoch_0" )} """.split() a : int = run_command(self._launch_args + testargs , return_stdout=__snake_case ) self.assertNotIn('epoch 0:' , __snake_case ) self.assertIn('epoch 1:' , __snake_case ) def lowercase_ ( self : int ): a : Optional[int] = F""" examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , "step_2" )} """.split() a : Union[str, Any] = run_command(self._launch_args + testargs , return_stdout=__snake_case ) if torch.cuda.is_available(): a : Any = torch.cuda.device_count() else: a : str = 1 if num_processes > 1: self.assertNotIn('epoch 0:' , __snake_case ) self.assertIn('epoch 1:' , __snake_case ) else: self.assertIn('epoch 0:' , __snake_case ) self.assertIn('epoch 1:' , __snake_case ) @slow def lowercase_ ( self : Tuple ): a : Tuple = '\n examples/by_feature/cross_validation.py\n --num_folds 2\n '.split() with mock.patch.dict(os.environ , {'TESTING_MOCKED_DATALOADERS': '0'} ): a : Any = run_command(self._launch_args + testargs , return_stdout=__snake_case ) a : Optional[Any] = re.findall('({.+})' , __snake_case ) a : str = [r for r in results if 'accuracy' in r][-1] a : str = ast.literal_eval(__snake_case ) self.assertGreaterEqual(results['accuracy'] , 0.75 ) def lowercase_ ( self : Optional[int] ): a : int = ['examples/by_feature/multi_process_metrics.py'] run_command(self._launch_args + testargs ) @require_trackers @mock.patch.dict(os.environ , {'WANDB_MODE': 'offline'} ) def lowercase_ ( self : Optional[int] ): with tempfile.TemporaryDirectory() as tmpdir: a : Optional[Any] = F""" examples/by_feature/tracking.py --with_tracking --project_dir {tmpdir} """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(__snake_case , 'tracking' ) ) ) def lowercase_ ( self : List[str] ): a : Optional[Any] = ['examples/by_feature/gradient_accumulation.py'] run_command(self._launch_args + testargs ) def lowercase_ ( self : int ): a : Optional[Any] = ['examples/by_feature/local_sgd.py'] run_command(self._launch_args + testargs )
297
1
"""simple docstring""" import numpy as np class lowerCamelCase_: '''simple docstring''' def __init__( self ): _lowerCamelCase = (0, 0) _lowerCamelCase = None _lowerCamelCase = 0 _lowerCamelCase = 0 _lowerCamelCase = 0 def __eq__( self , lowerCamelCase__ ): return self.position == cell.position def snake_case__ ( self ): print(self.position ) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__=(5, 5) ): _lowerCamelCase = np.zeros(__A ) _lowerCamelCase = world_size[0] _lowerCamelCase = world_size[1] def snake_case__ ( self ): print(self.w ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] _lowerCamelCase = cell.position[0] _lowerCamelCase = cell.position[1] _lowerCamelCase = [] for n in neughbour_cord: _lowerCamelCase = current_x + n[0] _lowerCamelCase = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: _lowerCamelCase = Cell() _lowerCamelCase = (x, y) _lowerCamelCase = cell neighbours.append(__A ) return neighbours def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Tuple , lowercase_ : Optional[Any] ) -> Tuple: _lowerCamelCase = [] _lowerCamelCase = [] _open.append(lowercase_ ) while _open: _lowerCamelCase = np.argmin([n.f for n in _open] ) _lowerCamelCase = _open[min_f] _closed.append(_open.pop(lowercase_ ) ) if current == goal: break for n in world.get_neigbours(lowercase_ ): for c in _closed: if c == n: continue _lowerCamelCase = current.g + 1 _lowerCamelCase = n.position _lowerCamelCase = goal.position _lowerCamelCase = (ya - ya) ** 2 + (xa - xa) ** 2 _lowerCamelCase = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(lowercase_ ) _lowerCamelCase = [] while current.parent is not None: path.append(current.position ) _lowerCamelCase = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Dict = Gridworld() # Start position and goal __SCREAMING_SNAKE_CASE : Optional[Any] = Cell() __SCREAMING_SNAKE_CASE : Dict = (0, 0) __SCREAMING_SNAKE_CASE : str = Cell() __SCREAMING_SNAKE_CASE : int = (4, 4) print(F"""path from {start.position} to {goal.position}""") __SCREAMING_SNAKE_CASE : Optional[int] = astar(world, start, goal) # Just for visual reasons. for i in s: __SCREAMING_SNAKE_CASE : Optional[Any] = 1 print(world.w)
360
"""simple docstring""" from __future__ import annotations def lowerCAmelCase_( lowercase_ : list , lowercase_ : int ) -> Tuple: # Checks if the entire collection has been sorted if len(lowercase_ ) <= 1 or n <= 1: return insert_next(lowercase_ , n - 1 ) rec_insertion_sort(lowercase_ , n - 1 ) def lowerCAmelCase_( lowercase_ : list , lowercase_ : int ) -> Any: # Checks order between adjacent elements if index >= len(lowercase_ ) or collection[index - 1] <= collection[index]: return # Swaps adjacent elements since they are not in ascending order _lowerCamelCase , _lowerCamelCase = ( collection[index], collection[index - 1], ) insert_next(lowercase_ , index + 1 ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Optional[int] = input('''Enter integers separated by spaces: ''') __SCREAMING_SNAKE_CASE : list[int] = [int(num) for num in numbers.split()] rec_insertion_sort(number_list, len(number_list)) print(number_list)
73
0
'''simple docstring''' import random def snake_case_ (_a : Dict , _a : Tuple , _a : List[str] ): UpperCAmelCase = a[left_index] UpperCAmelCase = left_index + 1 for j in range(left_index + 1 , lowerCAmelCase__ ): if a[j] < pivot: UpperCAmelCase = a[i], a[j] i += 1 UpperCAmelCase = a[i - 1], a[left_index] return i - 1 def snake_case_ (_a : Union[str, Any] , _a : Dict , _a : Dict ): if left < right: UpperCAmelCase = random.randint(lowerCAmelCase__ , right - 1 ) UpperCAmelCase = ( a[left], a[pivot], ) # switches the pivot with the left most bound UpperCAmelCase = partition(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) quick_sort_random( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) # recursive quicksort to the left of the pivot point quick_sort_random( lowerCAmelCase__ , pivot_index + 1 , lowerCAmelCase__ ) # recursive quicksort to the right of the pivot point def snake_case_ (): UpperCAmelCase = input('''Enter numbers separated by a comma:\n''' ).strip() UpperCAmelCase = [int(lowerCAmelCase__ ) for item in user_input.split(''',''' )] quick_sort_random(lowerCAmelCase__ , 0 , len(lowerCAmelCase__ ) ) print(lowerCAmelCase__ ) if __name__ == "__main__": main()
34
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _UpperCamelCase = { '''configuration_bridgetower''': [ '''BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BridgeTowerConfig''', '''BridgeTowerTextConfig''', '''BridgeTowerVisionConfig''', ], '''processing_bridgetower''': ['''BridgeTowerProcessor'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = ['''BridgeTowerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase = [ '''BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BridgeTowerForContrastiveLearning''', '''BridgeTowerForImageAndTextRetrieval''', '''BridgeTowerForMaskedLM''', '''BridgeTowerModel''', '''BridgeTowerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys _UpperCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
254
0
'''simple docstring''' import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 snake_case__ : Any = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') snake_case__ : List[str] = get_tests_dir('''fixtures/vocab.json''') snake_case__ : Dict = get_tests_dir('''fixtures''') class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' lowerCamelCase_ :Dict = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = 0 def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) def _UpperCamelCase ( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase_ : Optional[int] = WavaVecaConfig() UpperCAmelCase_ : int = AutoProcessor.from_pretrained('facebook/wav2vec2-base-960h' ) # save in new folder model_config.save_pretrained(lowerCamelCase_ ) processor.save_pretrained(lowerCamelCase_ ) UpperCAmelCase_ : Any = AutoProcessor.from_pretrained(lowerCamelCase_ ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) def _UpperCamelCase ( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(lowerCamelCase_ , os.path.join(lowerCamelCase_ , lowerCamelCase_ ) ) copyfile(lowerCamelCase_ , os.path.join(lowerCamelCase_ , 'vocab.json' ) ) UpperCAmelCase_ : Optional[Any] = AutoProcessor.from_pretrained(lowerCamelCase_ ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) def _UpperCamelCase ( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase_ : Optional[Any] = WavaVecaFeatureExtractor() UpperCAmelCase_ : Optional[int] = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) UpperCAmelCase_ : Tuple = WavaVecaProcessor(lowerCamelCase_ , lowerCamelCase_ ) # save in new folder processor.save_pretrained(lowerCamelCase_ ) # drop `processor_class` in tokenizer with open(os.path.join(lowerCamelCase_ , lowerCamelCase_ ) , 'r' ) as f: UpperCAmelCase_ : str = json.load(lowerCamelCase_ ) config_dict.pop('processor_class' ) with open(os.path.join(lowerCamelCase_ , lowerCamelCase_ ) , 'w' ) as f: f.write(json.dumps(lowerCamelCase_ ) ) UpperCAmelCase_ : Optional[int] = AutoProcessor.from_pretrained(lowerCamelCase_ ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) def _UpperCamelCase ( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase_ : int = WavaVecaFeatureExtractor() UpperCAmelCase_ : List[str] = AutoTokenizer.from_pretrained('facebook/wav2vec2-base-960h' ) UpperCAmelCase_ : Tuple = WavaVecaProcessor(lowerCamelCase_ , lowerCamelCase_ ) # save in new folder processor.save_pretrained(lowerCamelCase_ ) # drop `processor_class` in feature extractor with open(os.path.join(lowerCamelCase_ , lowerCamelCase_ ) , 'r' ) as f: UpperCAmelCase_ : Optional[int] = json.load(lowerCamelCase_ ) config_dict.pop('processor_class' ) with open(os.path.join(lowerCamelCase_ , lowerCamelCase_ ) , 'w' ) as f: f.write(json.dumps(lowerCamelCase_ ) ) UpperCAmelCase_ : Dict = AutoProcessor.from_pretrained(lowerCamelCase_ ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) def _UpperCamelCase ( self ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmpdirname: UpperCAmelCase_ : Dict = WavaVecaConfig(processor_class='Wav2Vec2Processor' ) model_config.save_pretrained(lowerCamelCase_ ) # copy relevant files copyfile(lowerCamelCase_ , os.path.join(lowerCamelCase_ , 'vocab.json' ) ) # create emtpy sample processor with open(os.path.join(lowerCamelCase_ , lowerCamelCase_ ) , 'w' ) as f: f.write('{}' ) UpperCAmelCase_ : Tuple = AutoProcessor.from_pretrained(lowerCamelCase_ ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) def _UpperCamelCase ( self ): '''simple docstring''' with self.assertRaises(lowerCamelCase_ ): UpperCAmelCase_ : List[str] = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowerCamelCase_ ): UpperCAmelCase_ : List[Any] = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowerCamelCase_ ) UpperCAmelCase_ : List[str] = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowerCamelCase_ ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) UpperCAmelCase_ : int = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , 'NewFeatureExtractor' ) UpperCAmelCase_ : Tuple = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizerFast' ) # Test we can also load the slow version UpperCAmelCase_ : List[Any] = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowerCamelCase_ , use_fast=lowerCamelCase_ ) UpperCAmelCase_ : int = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , 'NewTokenizer' ) else: self.assertEqual(tokenizer.__class__.__name__ , 'NewTokenizer' ) def _UpperCamelCase ( self ): '''simple docstring''' try: AutoConfig.register('custom' , lowerCamelCase_ ) AutoFeatureExtractor.register(lowerCamelCase_ , lowerCamelCase_ ) AutoTokenizer.register(lowerCamelCase_ , slow_tokenizer_class=lowerCamelCase_ ) AutoProcessor.register(lowerCamelCase_ , lowerCamelCase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCamelCase_ ): AutoProcessor.register(lowerCamelCase_ , lowerCamelCase_ ) # Now that the config is registered, it can be used as any other config with the auto-API UpperCAmelCase_ : Tuple = CustomFeatureExtractor.from_pretrained(lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase_ : Tuple = os.path.join(lowerCamelCase_ , 'vocab.txt' ) with open(lowerCamelCase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase_ : Dict = CustomTokenizer(lowerCamelCase_ ) UpperCAmelCase_ : Dict = CustomProcessor(lowerCamelCase_ , lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(lowerCamelCase_ ) UpperCAmelCase_ : Optional[Any] = AutoProcessor.from_pretrained(lowerCamelCase_ ) self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCamelCase ( self ): '''simple docstring''' class __SCREAMING_SNAKE_CASE ( __snake_case ): '''simple docstring''' lowerCamelCase_ :List[Any] = False class __SCREAMING_SNAKE_CASE ( __snake_case ): '''simple docstring''' lowerCamelCase_ :Optional[int] = False class __SCREAMING_SNAKE_CASE ( __snake_case ): '''simple docstring''' lowerCamelCase_ :Tuple = "AutoFeatureExtractor" lowerCamelCase_ :Any = "AutoTokenizer" lowerCamelCase_ :str = False try: AutoConfig.register('custom' , lowerCamelCase_ ) AutoFeatureExtractor.register(lowerCamelCase_ , lowerCamelCase_ ) AutoTokenizer.register(lowerCamelCase_ , slow_tokenizer_class=lowerCamelCase_ ) AutoProcessor.register(lowerCamelCase_ , lowerCamelCase_ ) # If remote code is not set, the default is to use local classes. UpperCAmelCase_ : Dict = AutoProcessor.from_pretrained('hf-internal-testing/test_dynamic_processor' ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. UpperCAmelCase_ : List[Any] = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowerCamelCase_ ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. UpperCAmelCase_ : List[Any] = AutoProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_processor' , trust_remote_code=lowerCamelCase_ ) self.assertEqual(processor.__class__.__name__ , 'NewProcessor' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-bert' ) self.assertEqual(processor.__class__.__name__ , 'BertTokenizerFast' ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Tuple = AutoProcessor.from_pretrained('hf-internal-testing/tiny-random-convnext' ) self.assertEqual(processor.__class__.__name__ , 'ConvNextImageProcessor' ) @is_staging_test class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' lowerCamelCase_ :Any = ["[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "bla", "blou"] @classmethod def _UpperCamelCase ( cls ): '''simple docstring''' UpperCAmelCase_ : Dict = TOKEN HfFolder.save_token(lowerCamelCase_ ) @classmethod def _UpperCamelCase ( cls ): '''simple docstring''' try: delete_repo(token=cls._token , repo_id='test-processor' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='valid_org/test-processor-org' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='test-dynamic-processor' ) except HTTPError: pass def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = WavaVecaProcessor.from_pretrained(lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(lowerCamelCase_ , 'test-processor' ) , push_to_hub=lowerCamelCase_ , use_auth_token=self._token ) UpperCAmelCase_ : Tuple = WavaVecaProcessor.from_pretrained(F'''{USER}/test-processor''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(lowerCamelCase_ , getattr(new_processor.feature_extractor , lowerCamelCase_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _UpperCamelCase ( self ): '''simple docstring''' UpperCAmelCase_ : Union[str, Any] = WavaVecaProcessor.from_pretrained(lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(lowerCamelCase_ , 'test-processor-org' ) , push_to_hub=lowerCamelCase_ , use_auth_token=self._token , organization='valid_org' , ) UpperCAmelCase_ : Any = WavaVecaProcessor.from_pretrained('valid_org/test-processor-org' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(lowerCamelCase_ , getattr(new_processor.feature_extractor , lowerCamelCase_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def _UpperCamelCase ( self ): '''simple docstring''' CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() UpperCAmelCase_ : Optional[int] = CustomFeatureExtractor.from_pretrained(lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: UpperCAmelCase_ : str = os.path.join(lowerCamelCase_ , 'vocab.txt' ) with open(lowerCamelCase_ , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in self.vocab_tokens] ) ) UpperCAmelCase_ : Union[str, Any] = CustomTokenizer(lowerCamelCase_ ) UpperCAmelCase_ : List[str] = CustomProcessor(lowerCamelCase_ , lowerCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(F'''{USER}/test-dynamic-processor''' , token=self._token ) UpperCAmelCase_ : str = Repository(lowerCamelCase_ , clone_from=F'''{USER}/test-dynamic-processor''' , token=self._token ) processor.save_pretrained(lowerCamelCase_ ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { 'AutoFeatureExtractor': 'custom_feature_extraction.CustomFeatureExtractor', 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(lowerCamelCase_ , 'tokenizer_config.json' ) ) as f: UpperCAmelCase_ : Any = json.load(lowerCamelCase_ ) self.assertDictEqual( tokenizer_config['auto_map'] , { 'AutoTokenizer': ['custom_tokenization.CustomTokenizer', None], 'AutoProcessor': 'custom_processing.CustomProcessor', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(lowerCamelCase_ , 'custom_feature_extraction.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(lowerCamelCase_ , 'custom_tokenization.py' ) ) ) self.assertTrue(os.path.isfile(os.path.join(lowerCamelCase_ , 'custom_processing.py' ) ) ) repo.push_to_hub() UpperCAmelCase_ : str = AutoProcessor.from_pretrained(F'''{USER}/test-dynamic-processor''' , trust_remote_code=lowerCamelCase_ ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , 'CustomProcessor' )
365
'''simple docstring''' snake_case__ : Optional[Any] = tuple[float, float, float] snake_case__ : Tuple = tuple[float, float, float] def _lowerCamelCase ( lowerCamelCase_ : Pointad , lowerCamelCase_ : Pointad ): """simple docstring""" UpperCAmelCase_ : Any = end_pointa[0] - end_pointa[0] UpperCAmelCase_ : Optional[Any] = end_pointa[1] - end_pointa[1] UpperCAmelCase_ : Any = end_pointa[2] - end_pointa[2] return (x, y, z) def _lowerCamelCase ( lowerCamelCase_ : Vectorad , lowerCamelCase_ : Vectorad ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = ab[1] * ac[2] - ab[2] * ac[1] # *i UpperCAmelCase_ : Optional[Any] = (ab[0] * ac[2] - ab[2] * ac[0]) * -1 # *j UpperCAmelCase_ : Dict = ab[0] * ac[1] - ab[1] * ac[0] # *k return (x, y, z) def _lowerCamelCase ( lowerCamelCase_ : Vectorad , lowerCamelCase_ : int ): """simple docstring""" return tuple(round(lowerCamelCase_ , lowerCamelCase_ ) for x in vector ) == (0, 0, 0) def _lowerCamelCase ( lowerCamelCase_ : Pointad , lowerCamelCase_ : Pointad , lowerCamelCase_ : Pointad , lowerCamelCase_ : int = 10 ): """simple docstring""" UpperCAmelCase_ : List[str] = create_vector(lowerCamelCase_ , lowerCamelCase_ ) UpperCAmelCase_ : List[Any] = create_vector(lowerCamelCase_ , lowerCamelCase_ ) return is_zero_vector(get_ad_vectors_cross(lowerCamelCase_ , lowerCamelCase_ ) , lowerCamelCase_ )
274
0
def a ( A__ : str , A__ : bool = False ) -> str: """simple docstring""" if not isinstance(A__ , A__ ): _lowercase =F'''Expected string as input, found {type(A__ )}''' raise ValueError(A__ ) if not isinstance(A__ , A__ ): _lowercase =F'''Expected boolean as use_pascal parameter, found {type(A__ )}''' raise ValueError(A__ ) _lowercase =input_str.split('_' ) _lowercase =0 if use_pascal else 1 _lowercase =words[start_index:] _lowercase =[word[0].upper() + word[1:] for word in words_to_capitalize] _lowercase ='' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
205
import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 lowercase_ = data_utils.TransfoXLTokenizer lowercase_ = data_utils.TransfoXLCorpus lowercase_ = data_utils lowercase_ = data_utils def a ( A__ : int , A__ : Dict , A__ : Union[str, Any] , A__ : Union[str, Any] ) -> List[str]: """simple docstring""" if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(A__ , 'rb' ) as fp: _lowercase =pickle.load(A__ , encoding='latin1' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) _lowercase =pytorch_dump_folder_path + '/' + VOCAB_FILES_NAMES['pretrained_vocab_file'] print(F'''Save vocabulary to {pytorch_vocab_dump_path}''' ) _lowercase =corpus.vocab.__dict__ torch.save(A__ , A__ ) _lowercase =corpus.__dict__ corpus_dict_no_vocab.pop('vocab' , A__ ) _lowercase =pytorch_dump_folder_path + '/' + CORPUS_NAME print(F'''Save dataset to {pytorch_dataset_dump_path}''' ) torch.save(A__ , A__ ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model _lowercase =os.path.abspath(A__ ) _lowercase =os.path.abspath(A__ ) print(F'''Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.''' ) # Initialise PyTorch model if transfo_xl_config_file == "": _lowercase =TransfoXLConfig() else: _lowercase =TransfoXLConfig.from_json_file(A__ ) print(F'''Building PyTorch model from configuration: {config}''' ) _lowercase =TransfoXLLMHeadModel(A__ ) _lowercase =load_tf_weights_in_transfo_xl(A__ , A__ , A__ ) # Save pytorch-model _lowercase =os.path.join(A__ , A__ ) _lowercase =os.path.join(A__ , A__ ) print(F'''Save PyTorch model to {os.path.abspath(A__ )}''' ) torch.save(model.state_dict() , A__ ) print(F'''Save configuration file to {os.path.abspath(A__ )}''' ) with open(A__ , 'w' , encoding='utf-8' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) lowercase_ = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
205
1
from functools import lru_cache @lru_cache def UpperCAmelCase_( a__ ): """simple docstring""" if num < 0: raise ValueError('''Number should not be negative.''' ) return 1 if num in (0, 1) else num * factorial(num - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
19
import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class a_ ( a__ ): """simple docstring""" def __init__( self , *_lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase=None , **_lowerCamelCase ) ->int: super().__init__(*_lowerCamelCase , **_lowerCamelCase ) SCREAMING_SNAKE_CASE : Dict = eval_examples SCREAMING_SNAKE_CASE : Optional[int] = post_process_function def __lowerCAmelCase ( self , _lowerCamelCase = None , _lowerCamelCase=None , _lowerCamelCase = None , _lowerCamelCase = "eval" , **_lowerCamelCase , ) ->Dict[str, float]: SCREAMING_SNAKE_CASE : Any = gen_kwargs.copy() SCREAMING_SNAKE_CASE : str = ( gen_kwargs['''max_length'''] if gen_kwargs.get('''max_length''' ) is not None else self.args.generation_max_length ) SCREAMING_SNAKE_CASE : Dict = ( gen_kwargs['''num_beams'''] if gen_kwargs.get('''num_beams''' ) is not None else self.args.generation_num_beams ) SCREAMING_SNAKE_CASE : Any = gen_kwargs SCREAMING_SNAKE_CASE : List[Any] = self.eval_dataset if eval_dataset is None else eval_dataset SCREAMING_SNAKE_CASE : str = self.get_eval_dataloader(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[str] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE : Optional[Any] = self.compute_metrics SCREAMING_SNAKE_CASE : str = None SCREAMING_SNAKE_CASE : Optional[Any] = time.time() SCREAMING_SNAKE_CASE : List[str] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE : Tuple = eval_loop( _lowerCamelCase , description='''Evaluation''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_lowerCamelCase , metric_key_prefix=_lowerCamelCase , ) finally: SCREAMING_SNAKE_CASE : Dict = compute_metrics SCREAMING_SNAKE_CASE : Tuple = self.args.eval_batch_size * self.args.world_size if F"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[F"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _lowerCamelCase , _lowerCamelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default SCREAMING_SNAKE_CASE : Tuple = self.post_process_function(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) SCREAMING_SNAKE_CASE : Optional[int] = self.compute_metrics(_lowerCamelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): SCREAMING_SNAKE_CASE : Optional[int] = metrics.pop(_lowerCamelCase ) metrics.update(output.metrics ) else: SCREAMING_SNAKE_CASE : List[Any] = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(_lowerCamelCase ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) SCREAMING_SNAKE_CASE : int = self.callback_handler.on_evaluate(self.args , self.state , self.control , _lowerCamelCase ) return metrics def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=None , _lowerCamelCase = "test" , **_lowerCamelCase ) ->int: SCREAMING_SNAKE_CASE : str = gen_kwargs.copy() SCREAMING_SNAKE_CASE : str = self.get_test_dataloader(_lowerCamelCase ) # Temporarily disable metric computation, we will do it in the loop here. SCREAMING_SNAKE_CASE : Dict = self.compute_metrics SCREAMING_SNAKE_CASE : Tuple = None SCREAMING_SNAKE_CASE : List[str] = time.time() SCREAMING_SNAKE_CASE : Optional[Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: SCREAMING_SNAKE_CASE : Any = eval_loop( _lowerCamelCase , description='''Prediction''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_lowerCamelCase , metric_key_prefix=_lowerCamelCase , ) finally: SCREAMING_SNAKE_CASE : Optional[int] = compute_metrics SCREAMING_SNAKE_CASE : List[Any] = self.args.eval_batch_size * self.args.world_size if F"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[F"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( _lowerCamelCase , _lowerCamelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output SCREAMING_SNAKE_CASE : Tuple = self.post_process_function(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , '''predict''' ) SCREAMING_SNAKE_CASE : Dict = self.compute_metrics(_lowerCamelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): SCREAMING_SNAKE_CASE : List[Any] = metrics.pop(_lowerCamelCase ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_lowerCamelCase )
19
1
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> None: lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = analyze_text(SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : List[Any] = list(' ' + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase__ : Tuple = sum(single_char_strings.values() ) # one length string lowerCAmelCase__ : Optional[int] = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase__ : Optional[int] = single_char_strings[ch] lowerCAmelCase__ : Dict = my_str / all_sum my_fir_sum += prob * math.loga(SCREAMING_SNAKE_CASE_ ) # entropy formula. # print entropy print(F'''{round(-1 * my_fir_sum ):.1f}''' ) # two len string lowerCAmelCase__ : Tuple = sum(two_char_strings.values() ) lowerCAmelCase__ : Union[str, Any] = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase__ : Union[str, Any] = cha + cha if sequence in two_char_strings: lowerCAmelCase__ : Optional[Any] = two_char_strings[sequence] lowerCAmelCase__ : Optional[int] = int(SCREAMING_SNAKE_CASE_ ) / all_sum my_sec_sum += prob * math.loga(SCREAMING_SNAKE_CASE_ ) # print second entropy print(F'''{round(-1 * my_sec_sum ):.1f}''' ) # print the difference between them print(F'''{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}''' ) def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ ) -> tuple[dict, dict]: lowerCAmelCase__ : int = Counter() # type: ignore lowerCAmelCase__ : Optional[int] = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(SCREAMING_SNAKE_CASE_ ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def lowerCAmelCase__ ( ) -> Optional[Any]: import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
212
from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING lowerCamelCase__ = logging.get_logger(__name__) @add_end_docstrings(__magic_name__ ) class A__ ( __magic_name__ ): def __init__( self : int , *a : Dict , **a : Union[str, Any] ): '''simple docstring''' super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _lowerCamelCase ( self : Dict , a : List[str]=None ): '''simple docstring''' lowerCAmelCase__ : Any = {} if top_k is not None: lowerCAmelCase__ : Tuple = top_k return {}, {}, postprocess_params def __call__( self : Any , a : Union[str, List[str], "Image.Image", List["Image.Image"]] , **a : List[Any] ): '''simple docstring''' return super().__call__(a , **a ) def _lowerCamelCase ( self : Any , a : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = load_image(a ) lowerCAmelCase__ : Optional[int] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _lowerCamelCase ( self : Optional[int] , a : List[str] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model(**a ) return model_outputs def _lowerCamelCase ( self : Optional[Any] , a : List[Any] , a : List[Any]=5 ): '''simple docstring''' if top_k > self.model.config.num_labels: lowerCAmelCase__ : Optional[int] = self.model.config.num_labels if self.framework == "pt": lowerCAmelCase__ : List[Any] = model_outputs.logits.softmax(-1 )[0] lowerCAmelCase__ , lowerCAmelCase__ : Optional[int] = probs.topk(a ) elif self.framework == "tf": lowerCAmelCase__ : Any = stable_softmax(model_outputs.logits , axis=-1 )[0] lowerCAmelCase__ : Any = tf.math.top_k(a , k=a ) lowerCAmelCase__ , lowerCAmelCase__ : int = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f'''Unsupported framework: {self.framework}''' ) lowerCAmelCase__ : List[Any] = scores.tolist() lowerCAmelCase__ : List[str] = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
212
1
'''simple docstring''' import warnings warnings.warn( 'memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: ' '`from accelerate import find_executable_batch_size` to avoid this warning.', FutureWarning, )
251
'''simple docstring''' import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def A_( A : List[Any]): UpperCamelCase = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(A , A) def A_( A : Any): UpperCamelCase = list(s_dict.keys()) for key in keys: if "transformer_layers" in key: UpperCamelCase = s_dict.pop(A) elif "subsample" in key: UpperCamelCase = s_dict.pop(A) def A_( A : Optional[int]): UpperCamelCase , UpperCamelCase = emb.weight.shape UpperCamelCase = nn.Linear(A , A , bias=A) UpperCamelCase = emb.weight.data return lin_layer def A_( A : Optional[int] , A : List[str]): UpperCamelCase = torch.load(A , map_location='cpu') UpperCamelCase = mam_aaa['args'] UpperCamelCase = mam_aaa['model'] UpperCamelCase = state_dict['decoder.output_projection.weight'] remove_ignore_keys_(A) rename_keys(A) UpperCamelCase = state_dict['decoder.embed_tokens.weight'].shape[0] UpperCamelCase = args.share_decoder_input_output_embed UpperCamelCase = [int(A) for i in args.conv_kernel_sizes.split(',')] UpperCamelCase = SpeechaTextConfig( vocab_size=A , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='relu' , num_conv_layers=len(A) , conv_channels=args.conv_channels , conv_kernel_sizes=A , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=A , num_beams=5 , max_length=200 , use_cache=A , decoder_start_token_id=2 , early_stopping=A , ) UpperCamelCase = SpeechaTextForConditionalGeneration(A) UpperCamelCase , UpperCamelCase = model.model.load_state_dict(A , strict=A) if len(A) > 0 and not set(A) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( 'Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,' f''' but all the following weights are missing {missing}''') if tie_embeds: UpperCamelCase = make_linear_from_emb(model.model.decoder.embed_tokens) else: UpperCamelCase = lm_head_weights model.save_pretrained(A) if __name__ == "__main__": lowerCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument('--fairseq_path', type=str, help='Path to the fairseq model (.pt) file.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') lowerCAmelCase : List[str] = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
251
1
'''simple docstring''' import argparse import OmegaConf import torch from diffusers import DDIMScheduler, LDMPipeline, UNetLDMModel, VQModel def UpperCamelCase( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): UpperCAmelCase : Union[str, Any] = OmegaConf.load(UpperCAmelCase_ ) UpperCAmelCase : int = torch.load(UpperCAmelCase_ , map_location='cpu' )['model'] UpperCAmelCase : Optional[Any] = list(state_dict.keys() ) # extract state_dict for VQVAE UpperCAmelCase : Any = {} UpperCAmelCase : Union[str, Any] = 'first_stage_model.' for key in keys: if key.startswith(UpperCAmelCase_ ): UpperCAmelCase : Tuple = state_dict[key] # extract state_dict for UNetLDM UpperCAmelCase : Dict = {} UpperCAmelCase : Optional[Any] = 'model.diffusion_model.' for key in keys: if key.startswith(UpperCAmelCase_ ): UpperCAmelCase : Union[str, Any] = state_dict[key] UpperCAmelCase : str = config.model.params.first_stage_config.params UpperCAmelCase : str = config.model.params.unet_config.params UpperCAmelCase : List[Any] = VQModel(**UpperCAmelCase_ ).eval() vqvae.load_state_dict(UpperCAmelCase_ ) UpperCAmelCase : List[str] = UNetLDMModel(**UpperCAmelCase_ ).eval() unet.load_state_dict(UpperCAmelCase_ ) UpperCAmelCase : List[Any] = DDIMScheduler( timesteps=config.model.params.timesteps , beta_schedule='scaled_linear' , beta_start=config.model.params.linear_start , beta_end=config.model.params.linear_end , clip_sample=UpperCAmelCase_ , ) UpperCAmelCase : Tuple = LDMPipeline(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) pipeline.save_pretrained(UpperCAmelCase_ ) if __name__ == "__main__": lowercase__ = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", type=str, required=True) parser.add_argument("--config_path", type=str, required=True) parser.add_argument("--output_path", type=str, required=True) lowercase__ = parser.parse_args() convert_ldm_original(args.checkpoint_path, args.config_path, args.output_path)
151
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. lowercase__ = {"LayoutLMv2Config", "LayoutLMv3Config"} @is_pipeline_test class A_ ( unittest.TestCase ): '''simple docstring''' UpperCAmelCase_ : Optional[Any] = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING UpperCAmelCase_ : int = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: UpperCAmelCase_ : Union[str, Any] = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: UpperCAmelCase_ : Optional[int] = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def UpperCAmelCase_ ( self : Any ) -> List[Any]: UpperCAmelCase : int = pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='pt' ) UpperCAmelCase : List[Any] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) UpperCAmelCase : int = text_classifier('This is great !' , top_k=2 ) self.assertEqual( nested_simplify(lowercase_ ) , [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}] ) UpperCAmelCase : str = text_classifier(['This is great !', 'This is bad'] , top_k=2 ) self.assertEqual( nested_simplify(lowercase_ ) , [ [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], ] , ) UpperCAmelCase : List[str] = text_classifier('This is great !' , top_k=1 ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) # Legacy behavior UpperCAmelCase : str = text_classifier('This is great !' , return_all_scores=lowercase_ ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) UpperCAmelCase : List[str] = text_classifier('This is great !' , return_all_scores=lowercase_ ) self.assertEqual( nested_simplify(lowercase_ ) , [[{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}]] ) UpperCAmelCase : Optional[int] = text_classifier(['This is great !', 'Something else'] , return_all_scores=lowercase_ ) self.assertEqual( nested_simplify(lowercase_ ) , [ [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], [{'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_1', 'score': 0.496}], ] , ) UpperCAmelCase : Any = text_classifier(['This is great !', 'Something else'] , return_all_scores=lowercase_ ) self.assertEqual( nested_simplify(lowercase_ ) , [ {'label': 'LABEL_0', 'score': 0.504}, {'label': 'LABEL_0', 'score': 0.504}, ] , ) @require_torch def UpperCAmelCase_ ( self : Optional[Any] ) -> Union[str, Any]: import torch UpperCAmelCase : str = pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='pt' , device=torch.device('cpu' ) , ) UpperCAmelCase : List[str] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) @require_tf def UpperCAmelCase_ ( self : Dict ) -> List[str]: UpperCAmelCase : List[Any] = pipeline( task='text-classification' , model='hf-internal-testing/tiny-random-distilbert' , framework='tf' ) UpperCAmelCase : Optional[int] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'LABEL_0', 'score': 0.504}] ) @slow @require_torch def UpperCAmelCase_ ( self : Any ) -> int: UpperCAmelCase : Optional[Any] = pipeline('text-classification' ) UpperCAmelCase : str = text_classifier('This is great !' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'POSITIVE', 'score': 1.0}] ) UpperCAmelCase : Any = text_classifier('This is bad !' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'NEGATIVE', 'score': 1.0}] ) UpperCAmelCase : Union[str, Any] = text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'POSITIVE', 'score': 0.988}] ) @slow @require_tf def UpperCAmelCase_ ( self : str ) -> List[Any]: UpperCAmelCase : Dict = pipeline('text-classification' , framework='tf' ) UpperCAmelCase : Union[str, Any] = text_classifier('This is great !' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'POSITIVE', 'score': 1.0}] ) UpperCAmelCase : Union[str, Any] = text_classifier('This is bad !' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'NEGATIVE', 'score': 1.0}] ) UpperCAmelCase : Dict = text_classifier('Birds are a type of animal' ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': 'POSITIVE', 'score': 0.988}] ) def UpperCAmelCase_ ( self : Dict , lowercase_ : List[Any] , lowercase_ : Optional[int] , lowercase_ : Any ) -> Union[str, Any]: UpperCAmelCase : List[Any] = TextClassificationPipeline(model=lowercase_ , tokenizer=lowercase_ ) return text_classifier, ["HuggingFace is in", "This is another test"] def UpperCAmelCase_ ( self : int , lowercase_ : str , lowercase_ : str ) -> List[str]: UpperCAmelCase : str = text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 UpperCAmelCase : Dict = 'HuggingFace is in' UpperCAmelCase : int = text_classifier(lowercase_ ) self.assertEqual(nested_simplify(lowercase_ ) , [{'label': ANY(lowercase_ ), 'score': ANY(lowercase_ )}] ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) UpperCAmelCase : Optional[Any] = ['HuggingFace is in ', 'Paris is in France'] UpperCAmelCase : Optional[Any] = text_classifier(lowercase_ ) self.assertEqual( nested_simplify(lowercase_ ) , [{'label': ANY(lowercase_ ), 'score': ANY(lowercase_ )}, {'label': ANY(lowercase_ ), 'score': ANY(lowercase_ )}] , ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() ) self.assertTrue(outputs[1]['label'] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format UpperCAmelCase : int = text_classifier(lowercase_ , top_k=lowercase_ ) UpperCAmelCase : Tuple = len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(lowercase_ ) , [[{'label': ANY(lowercase_ ), 'score': ANY(lowercase_ )}] * N, [{'label': ANY(lowercase_ ), 'score': ANY(lowercase_ )}] * N] , ) UpperCAmelCase : List[Any] = {'text': 'HuggingFace is in ', 'text_pair': 'Paris is in France'} UpperCAmelCase : Tuple = text_classifier(lowercase_ ) self.assertEqual( nested_simplify(lowercase_ ) , {'label': ANY(lowercase_ ), 'score': ANY(lowercase_ )} , ) self.assertTrue(outputs['label'] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. UpperCAmelCase : List[Any] = [['HuggingFace is in ', 'Paris is in France']] with self.assertRaises(lowercase_ ): text_classifier(lowercase_ ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility UpperCAmelCase : Dict = text_classifier([[['HuggingFace is in ', 'Paris is in France']]] ) self.assertEqual( nested_simplify(lowercase_ ) , [{'label': ANY(lowercase_ ), 'score': ANY(lowercase_ )}] , ) self.assertTrue(outputs[0]['label'] in model.config.idalabel.values() )
151
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType UpperCAmelCase__ = logging.get_logger(__name__) UpperCAmelCase__ = { """microsoft/layoutlmv3-base""": """https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json""", } class a ( lowerCAmelCase_ ): _snake_case : Any = 'layoutlmv3' def __init__( self : Optional[Any] , __lowerCAmelCase : Tuple=5_0265 , __lowerCAmelCase : Union[str, Any]=768 , __lowerCAmelCase : str=12 , __lowerCAmelCase : int=12 , __lowerCAmelCase : Any=3072 , __lowerCAmelCase : Any="gelu" , __lowerCAmelCase : Any=0.1 , __lowerCAmelCase : Tuple=0.1 , __lowerCAmelCase : Any=512 , __lowerCAmelCase : List[str]=2 , __lowerCAmelCase : Optional[Any]=0.02 , __lowerCAmelCase : Optional[int]=1e-5 , __lowerCAmelCase : int=1 , __lowerCAmelCase : Optional[Any]=0 , __lowerCAmelCase : Optional[Any]=2 , __lowerCAmelCase : List[str]=1024 , __lowerCAmelCase : Any=128 , __lowerCAmelCase : int=128 , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : List[Any]=32 , __lowerCAmelCase : Any=128 , __lowerCAmelCase : int=64 , __lowerCAmelCase : List[str]=256 , __lowerCAmelCase : Union[str, Any]=True , __lowerCAmelCase : Any=True , __lowerCAmelCase : List[str]=True , __lowerCAmelCase : Optional[Any]=224 , __lowerCAmelCase : List[Any]=3 , __lowerCAmelCase : int=16 , __lowerCAmelCase : Optional[Any]=None , **__lowerCAmelCase : Union[str, Any] , ): super().__init__( vocab_size=__lowerCAmelCase , hidden_size=__lowerCAmelCase , num_hidden_layers=__lowerCAmelCase , num_attention_heads=__lowerCAmelCase , intermediate_size=__lowerCAmelCase , hidden_act=__lowerCAmelCase , hidden_dropout_prob=__lowerCAmelCase , attention_probs_dropout_prob=__lowerCAmelCase , max_position_embeddings=__lowerCAmelCase , type_vocab_size=__lowerCAmelCase , initializer_range=__lowerCAmelCase , layer_norm_eps=__lowerCAmelCase , pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase , ) _UpperCAmelCase = max_ad_position_embeddings _UpperCAmelCase = coordinate_size _UpperCAmelCase = shape_size _UpperCAmelCase = has_relative_attention_bias _UpperCAmelCase = rel_pos_bins _UpperCAmelCase = max_rel_pos _UpperCAmelCase = has_spatial_attention_bias _UpperCAmelCase = rel_ad_pos_bins _UpperCAmelCase = max_rel_ad_pos _UpperCAmelCase = text_embed _UpperCAmelCase = visual_embed _UpperCAmelCase = input_size _UpperCAmelCase = num_channels _UpperCAmelCase = patch_size _UpperCAmelCase = classifier_dropout class a ( lowerCAmelCase_ ): _snake_case : str = version.parse('1.12' ) @property def lowerCAmelCase_ ( self : Dict ): # The order of inputs is different for question answering and sequence classification if self.task in ["question-answering", "sequence-classification"]: return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ("""attention_mask""", {0: """batch""", 1: """sequence"""}), ("""bbox""", {0: """batch""", 1: """sequence"""}), ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) else: return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ("""bbox""", {0: """batch""", 1: """sequence"""}), ("""attention_mask""", {0: """batch""", 1: """sequence"""}), ("""pixel_values""", {0: """batch""", 1: """num_channels"""}), ] ) @property def lowerCAmelCase_ ( self : List[Any] ): return 1e-5 @property def lowerCAmelCase_ ( self : List[str] ): return 12 def lowerCAmelCase_ ( self : str , __lowerCAmelCase : "ProcessorMixin" , __lowerCAmelCase : int = -1 , __lowerCAmelCase : int = -1 , __lowerCAmelCase : bool = False , __lowerCAmelCase : Optional["TensorType"] = None , __lowerCAmelCase : int = 3 , __lowerCAmelCase : int = 40 , __lowerCAmelCase : int = 40 , ): setattr(processor.image_processor , """apply_ocr""" , __lowerCAmelCase ) # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _UpperCAmelCase = compute_effective_axis_dimension( __lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _UpperCAmelCase = processor.tokenizer.num_special_tokens_to_add(__lowerCAmelCase ) _UpperCAmelCase = compute_effective_axis_dimension( __lowerCAmelCase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__lowerCAmelCase ) # Generate dummy inputs according to compute batch and sequence _UpperCAmelCase = [[""" """.join([processor.tokenizer.unk_token] ) * seq_length]] * batch_size # Generate dummy bounding boxes _UpperCAmelCase = [[[48, 84, 73, 128]]] * batch_size # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX # batch_size = compute_effective_axis_dimension(batch_size, fixed_dimension=OnnxConfig.default_fixed_batch) _UpperCAmelCase = self._generate_dummy_images(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCAmelCase = dict( processor( __lowerCAmelCase , text=__lowerCAmelCase , boxes=__lowerCAmelCase , return_tensors=__lowerCAmelCase , ) ) return inputs
30
"""simple docstring""" import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def __UpperCAmelCase ( lowercase=None ,lowercase=None ): """simple docstring""" return field(default_factory=lambda: default ,metadata=lowercase ) @dataclass class a : _snake_case : str = field( metadata={'help': 'The csv file to plot.'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={'help': 'Whether to plot along batch size or sequence length. Defaults to sequence length.'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={'help': 'Whether the csv file has time results or memory results. Defaults to memory results.'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={'help': 'Disable logarithmic scale when plotting'} , ) _snake_case : bool = field( default=lowerCAmelCase_ , metadata={ 'help': 'Whether the csv file has training results or inference results. Defaults to inference results.' } , ) _snake_case : Optional[str] = field( default=lowerCAmelCase_ , metadata={'help': 'Filename under which the plot will be saved. If unused no plot is saved.'} , ) _snake_case : Optional[List[str]] = list_field( default=lowerCAmelCase_ , metadata={'help': 'List of model names that are used instead of the ones in the csv file.'} ) def __UpperCAmelCase ( lowercase ): """simple docstring""" try: int(lowercase ) return True except ValueError: return False def __UpperCAmelCase ( lowercase ): """simple docstring""" try: float(lowercase ) return True except ValueError: return False class a : def __init__( self : int , __lowerCAmelCase : Union[str, Any] ): _UpperCAmelCase = args _UpperCAmelCase = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline="""""" ) as csv_file: _UpperCAmelCase = csv.DictReader(__lowerCAmelCase ) for row in reader: _UpperCAmelCase = row["""model"""] self.result_dict[model_name]["bsz"].append(int(row["""batch_size"""] ) ) self.result_dict[model_name]["seq_len"].append(int(row["""sequence_length"""] ) ) if can_convert_to_int(row["""result"""] ): # value is not None _UpperCAmelCase = int(row["""result"""] ) elif can_convert_to_float(row["""result"""] ): # value is not None _UpperCAmelCase = float(row["""result"""] ) def lowerCAmelCase_ ( self : Optional[Any] ): _UpperCAmelCase , _UpperCAmelCase = plt.subplots() _UpperCAmelCase = """Time usage""" if self.args.is_time else """Memory usage""" _UpperCAmelCase = title_str + """ for training""" if self.args.is_train else title_str + """ for inference""" if not self.args.no_log_scale: # set logarithm scales ax.set_xscale("""log""" ) ax.set_yscale("""log""" ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): _UpperCAmelCase = sorted(set(self.result_dict[model_name]["""bsz"""] ) ) _UpperCAmelCase = sorted(set(self.result_dict[model_name]["""seq_len"""] ) ) _UpperCAmelCase = self.result_dict[model_name]["""result"""] ((_UpperCAmelCase) , (_UpperCAmelCase)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) _UpperCAmelCase = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: _UpperCAmelCase = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=__lowerCAmelCase , ) else: _UpperCAmelCase = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((_UpperCAmelCase) , (_UpperCAmelCase)) = ( ("""batch_size""", """len""") if self.args.plot_along_batch else ("""in #tokens""", """bsz""") ) _UpperCAmelCase = np.asarray(__lowerCAmelCase , __lowerCAmelCase )[: len(__lowerCAmelCase )] plt.scatter( __lowerCAmelCase , __lowerCAmelCase , label=f'''{label_model_name} - {inner_loop_label}: {inner_loop_value}''' ) plt.plot(__lowerCAmelCase , __lowerCAmelCase , """--""" ) title_str += f''' {label_model_name} vs.''' _UpperCAmelCase = title_str[:-4] _UpperCAmelCase = """Time in s""" if self.args.is_time else """Memory in MB""" # plot plt.title(__lowerCAmelCase ) plt.xlabel(__lowerCAmelCase ) plt.ylabel(__lowerCAmelCase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def __UpperCAmelCase ( ): """simple docstring""" _UpperCAmelCase = HfArgumentParser(lowercase ) _UpperCAmelCase = parser.parse_args_into_dataclasses()[0] _UpperCAmelCase = Plot(args=lowercase ) plot.plot() if __name__ == "__main__": main()
30
1
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A : """simple docstring""" def __init__( self : Any,lowercase_ : List[str],lowercase_ : Any=1_3,lowercase_ : Union[str, Any]=3_2,lowercase_ : Dict=2,lowercase_ : str=3,lowercase_ : List[str]=1_6,lowercase_ : Optional[Any]=[1, 2, 1],lowercase_ : List[str]=[2, 2, 4],lowercase_ : Optional[int]=2,lowercase_ : str=2.0,lowercase_ : Optional[int]=True,lowercase_ : Any=0.0,lowercase_ : Optional[int]=0.0,lowercase_ : Union[str, Any]=0.1,lowercase_ : int="gelu",lowercase_ : List[Any]=False,lowercase_ : Any=True,lowercase_ : Any=0.02,lowercase_ : str=1E-5,lowercase_ : Tuple=True,lowercase_ : Dict=None,lowercase_ : int=True,lowercase_ : Union[str, Any]=1_0,lowercase_ : Tuple=8,)-> List[Any]: '''simple docstring''' A__ = parent A__ = batch_size A__ = image_size A__ = patch_size A__ = num_channels A__ = embed_dim A__ = depths A__ = num_heads A__ = window_size A__ = mlp_ratio A__ = qkv_bias A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = drop_path_rate A__ = hidden_act A__ = use_absolute_embeddings A__ = patch_norm A__ = layer_norm_eps A__ = initializer_range A__ = is_training A__ = scope A__ = use_labels A__ = type_sequence_label_size A__ = encoder_stride def snake_case__ ( self : Dict )-> Any: '''simple docstring''' A__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size],self.type_sequence_label_size ) A__ = self.get_config() return config, pixel_values, labels def snake_case__ ( self : Union[str, Any] )-> Tuple: '''simple docstring''' return SwinvaConfig( image_size=self.image_size,patch_size=self.patch_size,num_channels=self.num_channels,embed_dim=self.embed_dim,depths=self.depths,num_heads=self.num_heads,window_size=self.window_size,mlp_ratio=self.mlp_ratio,qkv_bias=self.qkv_bias,hidden_dropout_prob=self.hidden_dropout_prob,attention_probs_dropout_prob=self.attention_probs_dropout_prob,drop_path_rate=self.drop_path_rate,hidden_act=self.hidden_act,use_absolute_embeddings=self.use_absolute_embeddings,path_norm=self.patch_norm,layer_norm_eps=self.layer_norm_eps,initializer_range=self.initializer_range,encoder_stride=self.encoder_stride,) def snake_case__ ( self : str,lowercase_ : str,lowercase_ : Any,lowercase_ : int )-> Dict: '''simple docstring''' A__ = SwinvaModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_ ) A__ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) A__ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape,(self.batch_size, expected_seq_len, expected_dim) ) def snake_case__ ( self : Tuple,lowercase_ : int,lowercase_ : Dict,lowercase_ : List[str] )-> Optional[int]: '''simple docstring''' A__ = SwinvaForMaskedImageModeling(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_ ) self.parent.assertEqual( result.logits.shape,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images A__ = 1 A__ = SwinvaForMaskedImageModeling(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) A__ = model(lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, 1, self.image_size, self.image_size) ) def snake_case__ ( self : List[Any],lowercase_ : str,lowercase_ : Tuple,lowercase_ : Dict )-> Tuple: '''simple docstring''' A__ = self.type_sequence_label_size A__ = SwinvaForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.type_sequence_label_size) ) def snake_case__ ( self : int )-> str: '''simple docstring''' A__ = self.prepare_config_and_inputs() A__ , A__ , A__ = config_and_inputs A__ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowerCamelCase = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def snake_case__ ( self : Tuple )-> Tuple: '''simple docstring''' A__ = SwinvaModelTester(self ) A__ = ConfigTester(self,config_class=lowercase_,embed_dim=3_7 ) def snake_case__ ( self : Dict )-> Tuple: '''simple docstring''' self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self : int )-> Optional[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def snake_case__ ( self : int )-> Optional[int]: '''simple docstring''' pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def snake_case__ ( self : Any )-> Dict: '''simple docstring''' pass def snake_case__ ( self : int )-> List[Any]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(lowercase_ ) self.assertIsInstance(model.get_input_embeddings(),(nn.Module) ) A__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowercase_,nn.Linear ) ) def snake_case__ ( self : Any )-> int: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ = model_class(lowercase_ ) A__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ = [*signature.parameters.keys()] A__ = ['pixel_values'] self.assertListEqual(arg_names[:1],lowercase_ ) def snake_case__ ( self : List[str] )-> Tuple: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = True for model_class in self.all_model_classes: A__ = True A__ = False A__ = True A__ = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() with torch.no_grad(): A__ = model(**self._prepare_for_class(lowercase_,lowercase_ ) ) A__ = outputs.attentions A__ = len(self.model_tester.depths ) self.assertEqual(len(lowercase_ ),lowercase_ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] A__ = True A__ = config.window_size**2 A__ = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() with torch.no_grad(): A__ = model(**self._prepare_for_class(lowercase_,lowercase_ ) ) A__ = outputs.attentions self.assertEqual(len(lowercase_ ),lowercase_ ) self.assertListEqual( list(attentions[0].shape[-3:] ),[self.model_tester.num_heads[0], window_size_squared, window_size_squared],) A__ = len(lowercase_ ) # Check attention is always last and order is fine A__ = True A__ = True A__ = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() with torch.no_grad(): A__ = model(**self._prepare_for_class(lowercase_,lowercase_ ) ) if hasattr(self.model_tester,'num_hidden_states_types' ): A__ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states A__ = 2 self.assertEqual(out_len + added_hidden_states,len(lowercase_ ) ) A__ = outputs.attentions self.assertEqual(len(lowercase_ ),lowercase_ ) self.assertListEqual( list(self_attentions[0].shape[-3:] ),[self.model_tester.num_heads[0], window_size_squared, window_size_squared],) def snake_case__ ( self : Dict,lowercase_ : Tuple,lowercase_ : Optional[Any],lowercase_ : Optional[int],lowercase_ : Dict )-> List[Any]: '''simple docstring''' A__ = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() with torch.no_grad(): A__ = model(**self._prepare_for_class(lowercase_,lowercase_ ) ) A__ = outputs.hidden_states A__ = getattr( self.model_tester,'expected_num_hidden_layers',len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowercase_ ),lowercase_ ) # Swinv2 has a different seq_length A__ = ( config.patch_size if isinstance(config.patch_size,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) A__ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ),[num_patches, self.model_tester.embed_dim],) A__ = outputs.reshaped_hidden_states self.assertEqual(len(lowercase_ ),lowercase_ ) A__ , A__ , A__ , A__ = reshaped_hidden_states[0].shape A__ = ( reshaped_hidden_states[0].view(lowercase_,lowercase_,height * width ).permute(0,2,1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ),[num_patches, self.model_tester.embed_dim],) def snake_case__ ( self : int )-> List[Any]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: A__ = True self.check_hidden_states_output(lowercase_,lowercase_,lowercase_,lowercase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ = True self.check_hidden_states_output(lowercase_,lowercase_,lowercase_,lowercase_ ) def snake_case__ ( self : Union[str, Any] )-> Dict: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = 3 A__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) A__ = ( config.patch_size if isinstance(config.patch_size,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) A__ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) A__ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: A__ = True self.check_hidden_states_output(lowercase_,lowercase_,lowercase_,(padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ = True self.check_hidden_states_output(lowercase_,lowercase_,lowercase_,(padded_height, padded_width) ) def snake_case__ ( self : Optional[Any] )-> int: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowercase_ ) def snake_case__ ( self : Optional[int] )-> Tuple: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) @slow def snake_case__ ( self : str )-> Any: '''simple docstring''' for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ = SwinvaModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def snake_case__ ( self : Union[str, Any] )-> Tuple: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = _config_zero_init(lowercase_ ) for model_class in self.all_model_classes: A__ = model_class(config=lowercase_ ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item(),[0.0, 1.0],msg=F'Parameter {name} of model {model_class} seems not properly initialized',) @require_vision @require_torch class A ( unittest.TestCase ): """simple docstring""" @cached_property def snake_case__ ( self : int )-> Dict: '''simple docstring''' return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def snake_case__ ( self : List[Any] )-> Any: '''simple docstring''' A__ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( lowercase_ ) A__ = self.default_image_processor A__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) A__ = image_processor(images=lowercase_,return_tensors='pt' ).to(lowercase_ ) # forward pass with torch.no_grad(): A__ = model(**lowercase_ ) # verify the logits A__ = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape,lowercase_ ) A__ = torch.tensor([-0.3_947, -0.4_306, 0.0_026] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3],lowercase_,atol=1E-4 ) )
7
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class A : """simple docstring""" def __init__( self : str,lowercase_ : Any,lowercase_ : Tuple=1_3,lowercase_ : str=7,lowercase_ : Tuple=True,lowercase_ : int=True,lowercase_ : List[Any]=True,lowercase_ : List[str]=True,lowercase_ : List[str]=9_9,lowercase_ : List[Any]=6_4,lowercase_ : List[str]=5,lowercase_ : Optional[Any]=4,lowercase_ : Optional[Any]=3_7,lowercase_ : Optional[Any]="gelu",lowercase_ : int=0.1,lowercase_ : str=0.1,lowercase_ : Optional[Any]=5_1_2,lowercase_ : int=1_6,lowercase_ : List[Any]=2,lowercase_ : Union[str, Any]=0.02,lowercase_ : Tuple=3,lowercase_ : List[Any]=4,lowercase_ : str=None,)-> Union[str, Any]: '''simple docstring''' A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_input_mask A__ = use_token_type_ids A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = type_sequence_label_size A__ = initializer_range A__ = num_labels A__ = num_choices A__ = scope A__ = vocab_size - 1 def snake_case__ ( self : str )-> Optional[Any]: '''simple docstring''' A__ = ids_tensor([self.batch_size, self.seq_length],self.vocab_size ) A__ = None if self.use_input_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) A__ = None if self.use_labels: A__ = ids_tensor([self.batch_size, self.seq_length],self.num_labels ) A__ = self.get_config() return config, input_ids, input_mask, token_labels def snake_case__ ( self : List[Any] )-> Tuple: '''simple docstring''' return GPTNeoXConfig( vocab_size=self.vocab_size,hidden_size=self.hidden_size,num_hidden_layers=self.num_hidden_layers,num_attention_heads=self.num_attention_heads,intermediate_size=self.intermediate_size,hidden_act=self.hidden_act,hidden_dropout_prob=self.hidden_dropout_prob,attention_probs_dropout_prob=self.attention_probs_dropout_prob,max_position_embeddings=self.max_position_embeddings,type_vocab_size=self.type_vocab_size,is_decoder=lowercase_,initializer_range=self.initializer_range,pad_token_id=self.pad_token_id,) def snake_case__ ( self : Optional[int] )-> Union[str, Any]: '''simple docstring''' A__ , A__ , A__ , A__ = self.prepare_config_and_inputs() A__ = True return config, input_ids, input_mask, token_labels def snake_case__ ( self : Any,lowercase_ : List[Any],lowercase_ : List[Any],lowercase_ : str )-> Any: '''simple docstring''' A__ = GPTNeoXModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_ ) A__ = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape,(self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : Union[str, Any],lowercase_ : List[str],lowercase_ : Dict,lowercase_ : Optional[Any] )-> Tuple: '''simple docstring''' A__ = True A__ = GPTNeoXModel(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape,(self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self : Union[str, Any],lowercase_ : str,lowercase_ : Union[str, Any],lowercase_ : Union[str, Any],lowercase_ : List[str] )-> List[str]: '''simple docstring''' A__ = GPTNeoXForCausalLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self : Optional[int],lowercase_ : Optional[int],lowercase_ : Optional[int],lowercase_ : Dict,lowercase_ : Any )-> int: '''simple docstring''' A__ = self.num_labels A__ = GPTNeoXForQuestionAnswering(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_ ) self.parent.assertEqual(result.start_logits.shape,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape,(self.batch_size, self.seq_length) ) def snake_case__ ( self : List[str],lowercase_ : List[str],lowercase_ : int,lowercase_ : Union[str, Any],lowercase_ : Optional[int] )-> str: '''simple docstring''' A__ = self.num_labels A__ = GPTNeoXForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = ids_tensor([self.batch_size],self.type_sequence_label_size ) A__ = model(lowercase_,attention_mask=lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.num_labels) ) def snake_case__ ( self : Any,lowercase_ : Union[str, Any],lowercase_ : List[Any],lowercase_ : Optional[Any],lowercase_ : int )-> Union[str, Any]: '''simple docstring''' A__ = self.num_labels A__ = GPTNeoXForTokenClassification(lowercase_ ) model.to(lowercase_ ) model.eval() A__ = model(lowercase_,attention_mask=lowercase_,labels=lowercase_ ) self.parent.assertEqual(result.logits.shape,(self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self : int,lowercase_ : str,lowercase_ : int,lowercase_ : Union[str, Any] )-> List[Any]: '''simple docstring''' A__ = True A__ = GPTNeoXForCausalLM(config=lowercase_ ) model.to(lowercase_ ) model.eval() # first forward pass A__ = model(lowercase_,attention_mask=lowercase_,use_cache=lowercase_ ) A__ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A__ = ids_tensor((self.batch_size, 3),config.vocab_size ) A__ = ids_tensor((self.batch_size, 3),vocab_size=2 ) # append to next input_ids and A__ = torch.cat([input_ids, next_tokens],dim=-1 ) A__ = torch.cat([input_mask, next_mask],dim=-1 ) A__ = model(lowercase_,attention_mask=lowercase_,output_hidden_states=lowercase_ ) A__ = output_from_no_past['hidden_states'][0] A__ = model( lowercase_,attention_mask=lowercase_,past_key_values=lowercase_,output_hidden_states=lowercase_,)['hidden_states'][0] # select random slice A__ = ids_tensor((1,),output_from_past.shape[-1] ).item() A__ = output_from_no_past[:, -3:, random_slice_idx].detach() A__ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase_,lowercase_,atol=1E-3 ) ) def snake_case__ ( self : str )-> Union[str, Any]: '''simple docstring''' A__ = self.prepare_config_and_inputs() A__ , A__ , A__ , A__ = config_and_inputs A__ = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase = (GPTNeoXForCausalLM,) if is_torch_available() else () lowerCamelCase = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def snake_case__ ( self : str )-> Tuple: '''simple docstring''' A__ = GPTNeoXModelTester(self ) A__ = ConfigTester(self,config_class=lowercase_,hidden_size=6_4,num_attention_heads=8 ) def snake_case__ ( self : Optional[Any] )-> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def snake_case__ ( self : Union[str, Any] )-> Union[str, Any]: '''simple docstring''' A__ , A__ , A__ , A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowercase_,lowercase_,lowercase_ ) def snake_case__ ( self : Dict )-> List[Any]: '''simple docstring''' A__ , A__ , A__ , A__ = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowercase_,lowercase_,lowercase_ ) def snake_case__ ( self : List[str] )-> Any: '''simple docstring''' A__ , A__ , A__ , A__ = self.model_tester.prepare_config_and_inputs_for_decoder() A__ = None self.model_tester.create_and_check_model_as_decoder(lowercase_,lowercase_,lowercase_ ) def snake_case__ ( self : Optional[Any] )-> str: '''simple docstring''' A__ , A__ , A__ , A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowercase_,lowercase_,lowercase_ ) def snake_case__ ( self : Dict )-> Dict: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowercase_ ) def snake_case__ ( self : Tuple )-> List[Any]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase_ ) def snake_case__ ( self : Any )-> List[str]: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase_ ) def snake_case__ ( self : str )-> Tuple: '''simple docstring''' A__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase_ ) @unittest.skip(reason='Feed forward chunking is not implemented' ) def snake_case__ ( self : Union[str, Any] )-> Optional[Any]: '''simple docstring''' pass @parameterized.expand([('linear',), ('dynamic',)] ) def snake_case__ ( self : List[str],lowercase_ : Any )-> List[str]: '''simple docstring''' A__ , A__ = self.model_tester.prepare_config_and_inputs_for_common() A__ = ids_tensor([1, 1_0],config.vocab_size ) A__ = ids_tensor([1, int(config.max_position_embeddings * 1.5 )],config.vocab_size ) set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights A__ = GPTNeoXModel(lowercase_ ) original_model.to(lowercase_ ) original_model.eval() A__ = original_model(lowercase_ ).last_hidden_state A__ = original_model(lowercase_ ).last_hidden_state set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights A__ = {'type': scaling_type, 'factor': 10.0} A__ = GPTNeoXModel(lowercase_ ) scaled_model.to(lowercase_ ) scaled_model.eval() A__ = scaled_model(lowercase_ ).last_hidden_state A__ = scaled_model(lowercase_ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowercase_,lowercase_,atol=1E-5 ) ) else: self.assertFalse(torch.allclose(lowercase_,lowercase_,atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowercase_,lowercase_,atol=1E-5 ) ) @require_torch class A ( unittest.TestCase ): """simple docstring""" @slow def snake_case__ ( self : Tuple )-> Union[str, Any]: '''simple docstring''' A__ = AutoTokenizer.from_pretrained('EleutherAI/pythia-410m-deduped' ) for checkpointing in [True, False]: A__ = GPTNeoXForCausalLM.from_pretrained('EleutherAI/pythia-410m-deduped' ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(lowercase_ ) A__ = tokenizer('My favorite food is',return_tensors='pt' ).to(lowercase_ ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 A__ = 'My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI\'m not sure' A__ = model.generate(**lowercase_,do_sample=lowercase_,max_new_tokens=2_0 ) A__ = tokenizer.batch_decode(lowercase_ )[0] self.assertEqual(lowercase_,lowercase_ )
7
1
"""simple docstring""" print((lambda quine: quine % quine)("""print((lambda quine: quine %% quine)(%r))"""))
181
"""simple docstring""" # Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available snake_case_ = { """configuration_cpmant""": ["""CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CpmAntConfig"""], """tokenization_cpmant""": ["""CpmAntTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case_ = [ """CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST""", """CpmAntForCausalLM""", """CpmAntModel""", """CpmAntPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys snake_case_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
181
1
"""simple docstring""" from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TF_MODEL_FOR_MULTIPLE_CHOICE_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, ) if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Tuple, lowerCamelCase : Any, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : int=3, lowerCamelCase : int=4, lowerCamelCase : List[Any]=2, lowerCamelCase : Dict=7, lowerCamelCase : Dict=True, lowerCamelCase : Union[str, Any]=True, lowerCamelCase : Optional[int]=True, lowerCamelCase : Union[str, Any]=True, lowerCamelCase : Optional[int]=99, lowerCamelCase : List[Any]=36, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : List[str]=4, lowerCamelCase : List[str]=37, lowerCamelCase : List[Any]="gelu", lowerCamelCase : Any=0.1, lowerCamelCase : int=0.1, lowerCamelCase : Tuple=512, lowerCamelCase : int=16, lowerCamelCase : Union[str, Any]=2, lowerCamelCase : Union[str, Any]=0.02, lowerCamelCase : Union[str, Any]=6, lowerCamelCase : Tuple=6, lowerCamelCase : Any=3, lowerCamelCase : Optional[Any]=4, lowerCamelCase : Optional[Any]=None, lowerCamelCase : Tuple=1000, )-> str: lowerCamelCase__ : List[Any] =parent lowerCamelCase__ : Any =batch_size lowerCamelCase__ : Dict =num_channels lowerCamelCase__ : Optional[Any] =image_size lowerCamelCase__ : str =patch_size lowerCamelCase__ : int =is_training lowerCamelCase__ : Dict =use_input_mask lowerCamelCase__ : Tuple =use_token_type_ids lowerCamelCase__ : Dict =use_labels lowerCamelCase__ : List[str] =vocab_size lowerCamelCase__ : Optional[Any] =hidden_size lowerCamelCase__ : Tuple =num_hidden_layers lowerCamelCase__ : str =num_attention_heads lowerCamelCase__ : Optional[int] =intermediate_size lowerCamelCase__ : List[str] =hidden_act lowerCamelCase__ : str =hidden_dropout_prob lowerCamelCase__ : int =attention_probs_dropout_prob lowerCamelCase__ : Optional[Any] =max_position_embeddings lowerCamelCase__ : Union[str, Any] =type_vocab_size lowerCamelCase__ : int =type_sequence_label_size lowerCamelCase__ : Dict =initializer_range lowerCamelCase__ : str =coordinate_size lowerCamelCase__ : Tuple =shape_size lowerCamelCase__ : Optional[int] =num_labels lowerCamelCase__ : Optional[int] =num_choices lowerCamelCase__ : List[str] =scope lowerCamelCase__ : Dict =range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) lowerCamelCase__ : List[str] =text_seq_length lowerCamelCase__ : Any =(image_size // patch_size) ** 2 + 1 lowerCamelCase__ : Optional[Any] =self.text_seq_length + self.image_seq_length def snake_case ( self : Optional[Any] )-> int: lowerCamelCase__ : int =ids_tensor([self.batch_size, self.text_seq_length], self.vocab_size ) lowerCamelCase__ : List[Any] =ids_tensor([self.batch_size, self.text_seq_length, 4], self.range_bbox ) lowerCamelCase__ : str =bbox.numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowerCamelCase__ : Union[str, Any] =bbox[i, j, 3] lowerCamelCase__ : Dict =bbox[i, j, 1] lowerCamelCase__ : Tuple =tmp_coordinate if bbox[i, j, 2] < bbox[i, j, 0]: lowerCamelCase__ : str =bbox[i, j, 2] lowerCamelCase__ : List[Any] =bbox[i, j, 0] lowerCamelCase__ : Tuple =tmp_coordinate lowerCamelCase__ : List[str] =tf.constant(lowerCamelCase ) lowerCamelCase__ : str =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : List[str] =None if self.use_input_mask: lowerCamelCase__ : Union[str, Any] =random_attention_mask([self.batch_size, self.text_seq_length] ) lowerCamelCase__ : str =None if self.use_token_type_ids: lowerCamelCase__ : Optional[int] =ids_tensor([self.batch_size, self.text_seq_length], self.type_vocab_size ) lowerCamelCase__ : List[str] =None lowerCamelCase__ : int =None if self.use_labels: lowerCamelCase__ : Union[str, Any] =ids_tensor([self.batch_size], self.type_sequence_label_size ) lowerCamelCase__ : Optional[int] =ids_tensor([self.batch_size, self.text_seq_length], self.num_labels ) lowerCamelCase__ : Dict =LayoutLMvaConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, initializer_range=self.initializer_range, coordinate_size=self.coordinate_size, shape_size=self.shape_size, input_size=self.image_size, patch_size=self.patch_size, ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def snake_case ( self : int, lowerCamelCase : Optional[Any], lowerCamelCase : Any, lowerCamelCase : List[str], lowerCamelCase : Any, lowerCamelCase : Any, lowerCamelCase : Union[str, Any] )-> Dict: lowerCamelCase__ : str =TFLayoutLMvaModel(config=lowerCamelCase ) # text + image lowerCamelCase__ : Optional[Any] =model(lowerCamelCase, pixel_values=lowerCamelCase, training=lowerCamelCase ) lowerCamelCase__ : List[str] =model( lowerCamelCase, bbox=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, training=lowerCamelCase, ) lowerCamelCase__ : List[Any] =model(lowerCamelCase, bbox=lowerCamelCase, pixel_values=lowerCamelCase, training=lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) # text only lowerCamelCase__ : Any =model(lowerCamelCase, training=lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only lowerCamelCase__ : List[Any] =model({'''pixel_values''': pixel_values}, training=lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape, (self.batch_size, self.image_seq_length, self.hidden_size) ) def snake_case ( self : Optional[int], lowerCamelCase : Any, lowerCamelCase : List[str], lowerCamelCase : Any, lowerCamelCase : Dict, lowerCamelCase : int, lowerCamelCase : Any, lowerCamelCase : Any )-> List[Any]: lowerCamelCase__ : Dict =self.num_labels lowerCamelCase__ : Tuple =TFLayoutLMvaForSequenceClassification(config=lowerCamelCase ) lowerCamelCase__ : Any =model( lowerCamelCase, bbox=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, training=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def snake_case ( self : int, lowerCamelCase : Tuple, lowerCamelCase : Any, lowerCamelCase : Any, lowerCamelCase : str, lowerCamelCase : Optional[Any], lowerCamelCase : int, lowerCamelCase : Optional[int] )-> List[Any]: lowerCamelCase__ : str =self.num_labels lowerCamelCase__ : str =TFLayoutLMvaForTokenClassification(config=lowerCamelCase ) lowerCamelCase__ : Any =model( lowerCamelCase, bbox=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, labels=lowerCamelCase, training=lowerCamelCase, ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.text_seq_length, self.num_labels) ) def snake_case ( self : Union[str, Any], lowerCamelCase : int, lowerCamelCase : Dict, lowerCamelCase : Any, lowerCamelCase : List[str], lowerCamelCase : List[str], lowerCamelCase : int, lowerCamelCase : Tuple )-> Optional[Any]: lowerCamelCase__ : Any =2 lowerCamelCase__ : Optional[Any] =TFLayoutLMvaForQuestionAnswering(config=lowerCamelCase ) lowerCamelCase__ : Union[str, Any] =model( lowerCamelCase, bbox=lowerCamelCase, pixel_values=lowerCamelCase, attention_mask=lowerCamelCase, token_type_ids=lowerCamelCase, start_positions=lowerCamelCase, end_positions=lowerCamelCase, training=lowerCamelCase, ) self.parent.assertEqual(result.start_logits.shape, (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape, (self.batch_size, self.seq_length) ) def snake_case ( self : List[Any] )-> Optional[Any]: lowerCamelCase__ : Optional[Any] =self.prepare_config_and_inputs() ((lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__) , (lowerCamelCase__)) : Any =config_and_inputs lowerCamelCase__ : int ={ '''input_ids''': input_ids, '''bbox''': bbox, '''pixel_values''': pixel_values, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask, } return config, inputs_dict @require_tf class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' _a = ( ( TFLayoutLMvaModel, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, ) if is_tf_available() else () ) _a = ( {'document-question-answering': TFLayoutLMvaForQuestionAnswering, 'feature-extraction': TFLayoutLMvaModel} if is_tf_available() else {} ) _a = False _a = False _a = False def snake_case ( self : Dict, lowerCamelCase : int, lowerCamelCase : List[Any], lowerCamelCase : Optional[Any], lowerCamelCase : Optional[Any], lowerCamelCase : int )-> str: return True def snake_case ( self : List[str], lowerCamelCase : Tuple, lowerCamelCase : Tuple, lowerCamelCase : Optional[Any]=False )-> dict: lowerCamelCase__ : List[str] =copy.deepcopy(lowerCamelCase ) if model_class in get_values(lowerCamelCase ): lowerCamelCase__ : Dict ={ k: tf.tile(tf.expand_dims(lowerCamelCase, 1 ), (1, self.model_tester.num_choices) + (1,) * (v.ndim - 1) ) if isinstance(lowerCamelCase, tf.Tensor ) and v.ndim > 0 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(lowerCamelCase ): lowerCamelCase__ : Any =tf.ones(self.model_tester.batch_size, dtype=tf.intaa ) elif model_class in get_values(lowerCamelCase ): lowerCamelCase__ : Optional[Any] =tf.zeros(self.model_tester.batch_size, dtype=tf.intaa ) lowerCamelCase__ : Any =tf.zeros(self.model_tester.batch_size, dtype=tf.intaa ) elif model_class in get_values(lowerCamelCase ): lowerCamelCase__ : int =tf.zeros(self.model_tester.batch_size, dtype=tf.intaa ) elif model_class in get_values(lowerCamelCase ): lowerCamelCase__ : List[str] =tf.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length), dtype=tf.intaa ) return inputs_dict def snake_case ( self : str )-> Dict: lowerCamelCase__ : List[Any] =TFLayoutLMvaModelTester(self ) lowerCamelCase__ : List[str] =ConfigTester(self, config_class=lowerCamelCase, hidden_size=37 ) def snake_case ( self : Optional[int] )-> Optional[Any]: self.config_tester.run_common_tests() def snake_case ( self : Any )-> Optional[int]: lowerCamelCase__ , lowerCamelCase__ : Any =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Optional[int] =model_class(lowerCamelCase ) if getattr(lowerCamelCase, '''hf_compute_loss''', lowerCamelCase ): # The number of elements in the loss should be the same as the number of elements in the label lowerCamelCase__ : Dict =self._prepare_for_class(inputs_dict.copy(), lowerCamelCase, return_labels=lowerCamelCase ) lowerCamelCase__ : List[str] =prepared_for_class[ sorted(prepared_for_class.keys() - inputs_dict.keys(), reverse=lowerCamelCase )[0] ] lowerCamelCase__ : int =added_label.shape.as_list()[:1] # Test that model correctly compute the loss with kwargs lowerCamelCase__ : Union[str, Any] =self._prepare_for_class(inputs_dict.copy(), lowerCamelCase, return_labels=lowerCamelCase ) lowerCamelCase__ : List[str] =prepared_for_class.pop('''input_ids''' ) lowerCamelCase__ : Optional[Any] =model(lowerCamelCase, **lowerCamelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss when we mask some positions lowerCamelCase__ : Optional[Any] =self._prepare_for_class(inputs_dict.copy(), lowerCamelCase, return_labels=lowerCamelCase ) lowerCamelCase__ : Optional[int] =prepared_for_class.pop('''input_ids''' ) if "labels" in prepared_for_class: lowerCamelCase__ : Tuple =prepared_for_class['''labels'''].numpy() if len(labels.shape ) > 1 and labels.shape[1] != 1: lowerCamelCase__ : Union[str, Any] =-100 lowerCamelCase__ : Any =tf.convert_to_tensor(lowerCamelCase ) lowerCamelCase__ : Any =model(lowerCamelCase, **lowerCamelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) self.assertTrue(not np.any(np.isnan(loss.numpy() ) ) ) # Test that model correctly compute the loss with a dict lowerCamelCase__ : Union[str, Any] =self._prepare_for_class(inputs_dict.copy(), lowerCamelCase, return_labels=lowerCamelCase ) lowerCamelCase__ : Optional[int] =model(lowerCamelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss with a tuple lowerCamelCase__ : Optional[int] =self._prepare_for_class(inputs_dict.copy(), lowerCamelCase, return_labels=lowerCamelCase ) # Get keys that were added with the _prepare_for_class function lowerCamelCase__ : Optional[Any] =prepared_for_class.keys() - inputs_dict.keys() lowerCamelCase__ : List[str] =inspect.signature(model.call ).parameters lowerCamelCase__ : Union[str, Any] =list(signature.keys() ) # Create a dictionary holding the location of the tensors in the tuple lowerCamelCase__ : Optional[Any] ={0: '''input_ids'''} for label_key in label_keys: lowerCamelCase__ : str =signature_names.index(lowerCamelCase ) lowerCamelCase__ : Dict =label_key lowerCamelCase__ : Tuple =sorted(tuple_index_mapping.items() ) # Initialize a list with their default values, update the values and convert to a tuple lowerCamelCase__ : Optional[Any] =[] for name in signature_names: if name != "kwargs": list_input.append(signature[name].default ) for index, value in sorted_tuple_index_mapping: lowerCamelCase__ : Optional[Any] =prepared_for_class[value] lowerCamelCase__ : Tuple =tuple(lowerCamelCase ) # Send to model lowerCamelCase__ : Union[str, Any] =model(tuple_input[:-1] )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) def snake_case ( self : List[Any] )-> Optional[int]: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) def snake_case ( self : str )-> Any: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Any =self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowerCamelCase__ : Optional[Any] =type self.model_tester.create_and_check_model(lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) def snake_case ( self : Tuple )-> str: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) def snake_case ( self : Optional[int] )-> Optional[Any]: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : int =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) def snake_case ( self : Dict )-> Any: ( ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ( lowerCamelCase__ ) , ) : Optional[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ) @slow def snake_case ( self : Union[str, Any] )-> List[str]: for model_name in TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : List[Any] =TFLayoutLMvaModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def snake_case__ ( ): """simple docstring""" lowerCamelCase__ : List[str] =Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case ( self : List[str] )-> Tuple: return LayoutLMvaImageProcessor(apply_ocr=lowerCamelCase ) if is_vision_available() else None @slow def snake_case ( self : int )-> str: lowerCamelCase__ : Optional[Any] =TFLayoutLMvaModel.from_pretrained('''microsoft/layoutlmv3-base''' ) lowerCamelCase__ : Optional[int] =self.default_image_processor lowerCamelCase__ : Dict =prepare_img() lowerCamelCase__ : List[Any] =image_processor(images=lowerCamelCase, return_tensors='''tf''' ).pixel_values lowerCamelCase__ : str =tf.constant([[1, 2]] ) lowerCamelCase__ : Dict =tf.expand_dims(tf.constant([[1, 2, 3, 4], [5, 6, 7, 8]] ), axis=0 ) # forward pass lowerCamelCase__ : int =model(input_ids=lowerCamelCase, bbox=lowerCamelCase, pixel_values=lowerCamelCase, training=lowerCamelCase ) # verify the logits lowerCamelCase__ : str =(1, 199, 768) self.assertEqual(outputs.last_hidden_state.shape, lowerCamelCase ) lowerCamelCase__ : str =tf.constant( [[-0.0_529, 0.3_618, 0.1_632], [-0.1_587, -0.1_667, -0.0_400], [-0.1_557, -0.1_671, -0.0_505]] ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3], lowerCamelCase, atol=1E-4 ) )
238
"""simple docstring""" import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def __get__( self : Tuple, lowerCamelCase : List[str], lowerCamelCase : Optional[int]=None )-> List[str]: # See docs.python.org/3/howto/descriptor.html#properties if obj is None: return self if self.fget is None: raise AttributeError('''unreadable attribute''' ) lowerCamelCase__ : List[str] ='''__cached_''' + self.fget.__name__ lowerCamelCase__ : List[Any] =getattr(lowerCamelCase, lowerCamelCase, lowerCamelCase ) if cached is None: lowerCamelCase__ : Optional[int] =self.fget(lowerCamelCase ) setattr(lowerCamelCase, lowerCamelCase, lowerCamelCase ) return cached def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" lowerCamelCase__ : Optional[Any] =val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(f'''invalid truth value {val!r}''' ) def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" if is_torch_fx_proxy(__lowerCamelCase ): return True if is_torch_available(): import torch if isinstance(__lowerCamelCase , torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(__lowerCamelCase , tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(__lowerCamelCase , (jnp.ndarray, Tracer) ): return True return isinstance(__lowerCamelCase , np.ndarray ) def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" return isinstance(__lowerCamelCase , np.ndarray ) def snake_case__ ( __lowerCamelCase : Union[str, Any] ): """simple docstring""" return _is_numpy(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : Any ): """simple docstring""" import torch return isinstance(__lowerCamelCase , torch.Tensor ) def snake_case__ ( __lowerCamelCase : Optional[Any] ): """simple docstring""" return False if not is_torch_available() else _is_torch(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : List[str] ): """simple docstring""" import torch return isinstance(__lowerCamelCase , torch.device ) def snake_case__ ( __lowerCamelCase : Optional[Any] ): """simple docstring""" return False if not is_torch_available() else _is_torch_device(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : Tuple ): """simple docstring""" import torch if isinstance(__lowerCamelCase , __lowerCamelCase ): if hasattr(__lowerCamelCase , __lowerCamelCase ): lowerCamelCase__ : Tuple =getattr(__lowerCamelCase , __lowerCamelCase ) else: return False return isinstance(__lowerCamelCase , torch.dtype ) def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" return False if not is_torch_available() else _is_torch_dtype(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : Optional[Any] ): """simple docstring""" import tensorflow as tf return isinstance(__lowerCamelCase , tf.Tensor ) def snake_case__ ( __lowerCamelCase : int ): """simple docstring""" return False if not is_tf_available() else _is_tensorflow(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(__lowerCamelCase , '''is_symbolic_tensor''' ): return tf.is_symbolic_tensor(__lowerCamelCase ) return type(__lowerCamelCase ) == tf.Tensor def snake_case__ ( __lowerCamelCase : Optional[Any] ): """simple docstring""" return False if not is_tf_available() else _is_tf_symbolic_tensor(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" import jax.numpy as jnp # noqa: F811 return isinstance(__lowerCamelCase , jnp.ndarray ) def snake_case__ ( __lowerCamelCase : Tuple ): """simple docstring""" return False if not is_flax_available() else _is_jax(__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : List[str] ): """simple docstring""" if isinstance(__lowerCamelCase , (dict, UserDict) ): return {k: to_py_obj(__lowerCamelCase ) for k, v in obj.items()} elif isinstance(__lowerCamelCase , (list, tuple) ): return [to_py_obj(__lowerCamelCase ) for o in obj] elif is_tf_tensor(__lowerCamelCase ): return obj.numpy().tolist() elif is_torch_tensor(__lowerCamelCase ): return obj.detach().cpu().tolist() elif is_jax_tensor(__lowerCamelCase ): return np.asarray(__lowerCamelCase ).tolist() elif isinstance(__lowerCamelCase , (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" if isinstance(__lowerCamelCase , (dict, UserDict) ): return {k: to_numpy(__lowerCamelCase ) for k, v in obj.items()} elif isinstance(__lowerCamelCase , (list, tuple) ): return np.array(__lowerCamelCase ) elif is_tf_tensor(__lowerCamelCase ): return obj.numpy() elif is_torch_tensor(__lowerCamelCase ): return obj.detach().cpu().numpy() elif is_jax_tensor(__lowerCamelCase ): return np.asarray(__lowerCamelCase ) else: return obj class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' def snake_case ( self : int )-> Optional[int]: lowerCamelCase__ : Union[str, Any] =fields(self ) # Safety and consistency checks if not len(lowerCamelCase ): raise ValueError(F'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(F'''{self.__class__.__name__} should not have more than one required field.''' ) lowerCamelCase__ : List[Any] =getattr(self, class_fields[0].name ) lowerCamelCase__ : Union[str, Any] =all(getattr(self, field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(lowerCamelCase ): if isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase__ : Optional[int] =first_field.items() lowerCamelCase__ : Union[str, Any] =True else: try: lowerCamelCase__ : int =iter(lowerCamelCase ) lowerCamelCase__ : List[Any] =True except TypeError: lowerCamelCase__ : List[Any] =False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(lowerCamelCase ): if ( not isinstance(lowerCamelCase, (list, tuple) ) or not len(lowerCamelCase ) == 2 or not isinstance(element[0], lowerCamelCase ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute lowerCamelCase__ : Optional[int] =first_field else: # If we have a mixed iterator, raise an error raise ValueError( F'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self, element[0], element[1] ) if element[1] is not None: lowerCamelCase__ : str =element[1] elif first_field is not None: lowerCamelCase__ : Dict =first_field else: for field in class_fields: lowerCamelCase__ : Union[str, Any] =getattr(self, field.name ) if v is not None: lowerCamelCase__ : Optional[int] =v def __delitem__( self : int, *lowerCamelCase : List[str], **lowerCamelCase : Optional[int] )-> str: raise Exception(F'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def snake_case ( self : Optional[int], *lowerCamelCase : int, **lowerCamelCase : List[str] )-> Optional[Any]: raise Exception(F'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def snake_case ( self : Dict, *lowerCamelCase : Optional[int], **lowerCamelCase : Optional[Any] )-> int: raise Exception(F'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def snake_case ( self : List[Any], *lowerCamelCase : Tuple, **lowerCamelCase : List[Any] )-> Optional[int]: raise Exception(F'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Optional[Any], lowerCamelCase : Optional[int] )-> List[Any]: if isinstance(lowerCamelCase, lowerCamelCase ): lowerCamelCase__ : Union[str, Any] =dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : Union[str, Any], lowerCamelCase : List[str], lowerCamelCase : List[str] )-> Dict: if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(lowerCamelCase, lowerCamelCase ) super().__setattr__(lowerCamelCase, lowerCamelCase ) def __setitem__( self : Optional[Any], lowerCamelCase : Union[str, Any], lowerCamelCase : int )-> List[Any]: # Will raise a KeyException if needed super().__setitem__(lowerCamelCase, lowerCamelCase ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(lowerCamelCase, lowerCamelCase ) def snake_case ( self : str )-> Tuple[Any]: return tuple(self[k] for k in self.keys() ) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , lowerCAmelCase_ ): '''simple docstring''' @classmethod def snake_case ( cls : Optional[Any], lowerCamelCase : int )-> str: raise ValueError( F'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 'longest' _a = 'max_length' _a = 'do_not_pad' class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ): '''simple docstring''' _a = 'pt' _a = 'tf' _a = 'np' _a = 'jax' class __SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Optional[int], lowerCamelCase : List[ContextManager] )-> str: lowerCamelCase__ : List[str] =context_managers lowerCamelCase__ : int =ExitStack() def __enter__( self : List[str] )-> Union[str, Any]: for context_manager in self.context_managers: self.stack.enter_context(lowerCamelCase ) def __exit__( self : Tuple, *lowerCamelCase : Union[str, Any], **lowerCamelCase : Tuple )-> List[Any]: self.stack.__exit__(*lowerCamelCase, **lowerCamelCase ) def snake_case__ ( __lowerCamelCase : int ): """simple docstring""" lowerCamelCase__ : Tuple =infer_framework(__lowerCamelCase ) if framework == "tf": lowerCamelCase__ : Any =inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": lowerCamelCase__ : Tuple =inspect.signature(model_class.forward ) # PyTorch models else: lowerCamelCase__ : List[str] =inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def snake_case__ ( __lowerCamelCase : str ): """simple docstring""" lowerCamelCase__ : Optional[Any] =model_class.__name__ lowerCamelCase__ : Tuple =infer_framework(__lowerCamelCase ) if framework == "tf": lowerCamelCase__ : int =inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": lowerCamelCase__ : Any =inspect.signature(model_class.forward ) # PyTorch models else: lowerCamelCase__ : Union[str, Any] =inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def snake_case__ ( __lowerCamelCase : MutableMapping , __lowerCamelCase : str = "" , __lowerCamelCase : str = "." ): """simple docstring""" def _flatten_dict(__lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int]="" , __lowerCamelCase : str="." ): for k, v in d.items(): lowerCamelCase__ : List[str] =str(__lowerCamelCase ) + delimiter + str(__lowerCamelCase ) if parent_key else k if v and isinstance(__lowerCamelCase , __lowerCamelCase ): yield from flatten_dict(__lowerCamelCase , __lowerCamelCase , delimiter=__lowerCamelCase ).items() else: yield key, v return dict(_flatten_dict(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) ) @contextmanager def snake_case__ ( __lowerCamelCase : Tuple , __lowerCamelCase : bool = False ): """simple docstring""" if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def snake_case__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Dict=None ): """simple docstring""" if is_numpy_array(__lowerCamelCase ): return np.transpose(__lowerCamelCase , axes=__lowerCamelCase ) elif is_torch_tensor(__lowerCamelCase ): return array.T if axes is None else array.permute(*__lowerCamelCase ) elif is_tf_tensor(__lowerCamelCase ): import tensorflow as tf return tf.transpose(__lowerCamelCase , perm=__lowerCamelCase ) elif is_jax_tensor(__lowerCamelCase ): return jnp.transpose(__lowerCamelCase , axes=__lowerCamelCase ) else: raise ValueError(f'''Type not supported for transpose: {type(__lowerCamelCase )}.''' ) def snake_case__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[int] ): """simple docstring""" if is_numpy_array(__lowerCamelCase ): return np.reshape(__lowerCamelCase , __lowerCamelCase ) elif is_torch_tensor(__lowerCamelCase ): return array.reshape(*__lowerCamelCase ) elif is_tf_tensor(__lowerCamelCase ): import tensorflow as tf return tf.reshape(__lowerCamelCase , __lowerCamelCase ) elif is_jax_tensor(__lowerCamelCase ): return jnp.reshape(__lowerCamelCase , __lowerCamelCase ) else: raise ValueError(f'''Type not supported for reshape: {type(__lowerCamelCase )}.''' ) def snake_case__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : List[str]=None ): """simple docstring""" if is_numpy_array(__lowerCamelCase ): return np.squeeze(__lowerCamelCase , axis=__lowerCamelCase ) elif is_torch_tensor(__lowerCamelCase ): return array.squeeze() if axis is None else array.squeeze(dim=__lowerCamelCase ) elif is_tf_tensor(__lowerCamelCase ): import tensorflow as tf return tf.squeeze(__lowerCamelCase , axis=__lowerCamelCase ) elif is_jax_tensor(__lowerCamelCase ): return jnp.squeeze(__lowerCamelCase , axis=__lowerCamelCase ) else: raise ValueError(f'''Type not supported for squeeze: {type(__lowerCamelCase )}.''' ) def snake_case__ ( __lowerCamelCase : Dict , __lowerCamelCase : Optional[int] ): """simple docstring""" if is_numpy_array(__lowerCamelCase ): return np.expand_dims(__lowerCamelCase , __lowerCamelCase ) elif is_torch_tensor(__lowerCamelCase ): return array.unsqueeze(dim=__lowerCamelCase ) elif is_tf_tensor(__lowerCamelCase ): import tensorflow as tf return tf.expand_dims(__lowerCamelCase , axis=__lowerCamelCase ) elif is_jax_tensor(__lowerCamelCase ): return jnp.expand_dims(__lowerCamelCase , axis=__lowerCamelCase ) else: raise ValueError(f'''Type not supported for expand_dims: {type(__lowerCamelCase )}.''' ) def snake_case__ ( __lowerCamelCase : List[Any] ): """simple docstring""" if is_numpy_array(__lowerCamelCase ): return np.size(__lowerCamelCase ) elif is_torch_tensor(__lowerCamelCase ): return array.numel() elif is_tf_tensor(__lowerCamelCase ): import tensorflow as tf return tf.size(__lowerCamelCase ) elif is_jax_tensor(__lowerCamelCase ): return array.size else: raise ValueError(f'''Type not supported for expand_dims: {type(__lowerCamelCase )}.''' ) def snake_case__ ( __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple ): """simple docstring""" for key, value in auto_map.items(): if isinstance(__lowerCamelCase , (tuple, list) ): lowerCamelCase__ : Optional[int] =[f'''{repo_id}--{v}''' if (v is not None and '''--''' not in v) else v for v in value] elif value is not None and "--" not in value: lowerCamelCase__ : Tuple =f'''{repo_id}--{value}''' return auto_map def snake_case__ ( __lowerCamelCase : Optional[int] ): """simple docstring""" for base_class in inspect.getmro(__lowerCamelCase ): lowerCamelCase__ : Tuple =base_class.__module__ lowerCamelCase__ : Tuple =base_class.__name__ if module.startswith('''tensorflow''' ) or module.startswith('''keras''' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('''torch''' ) or name == "PreTrainedModel": return "pt" elif module.startswith('''flax''' ) or module.startswith('''jax''' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(f'''Could not infer framework from class {model_class}.''' )
238
1
import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _lowerCamelCase( _a, unittest.TestCase ): # TODO: is there an appropriate internal test set? lowercase_ : Optional[int] = """ssube/stable-diffusion-x4-upscaler-onnx""" def UpperCamelCase ( self, lowerCamelCase=0) -> str: """simple docstring""" _lowercase : List[Any] = floats_tensor((1, 3, 1_28, 1_28), rng=random.Random(lowerCamelCase)) _lowercase : List[str] = torch.manual_seed(lowerCamelCase) _lowercase : Any = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'output_type': 'numpy', } return inputs def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = self.get_dummy_inputs() _lowercase : Tuple = pipe(**lowerCamelCase).images _lowercase : Optional[Any] = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3]) assert np.abs(image_slice - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : List[str] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Optional[Any] = PNDMScheduler.from_config(pipe.scheduler.config, skip_prk_steps=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = self.get_dummy_inputs() _lowercase : Tuple = pipe(**lowerCamelCase).images _lowercase : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : int = np.array( [0.6_8_9_8_8_9_2, 0.5_9_2_4_0_5_5_6, 0.5_2_4_9_9_5_2_7, 0.5_8_8_6_6_2_1_5, 0.5_2_2_5_8_2_3_5, 0.5_2_5_7_2_7_1_5, 0.6_2_4_1_4_4_7_3, 0.6_1_7_4_3_8_7, 0.6_2_1_4_9_6_4]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Any = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : int = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : str = self.get_dummy_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images _lowercase : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Union[str, Any] = np.array( [0.7_6_5_9_2_7_8, 0.7_6_4_3_7_6_6_4, 0.7_5_5_7_9_1_0_7, 0.7_6_9_1_1_1_6, 0.7_7_6_6_6_9_8_6, 0.7_7_2_7_6_7_2, 0.7_7_5_8_6_6_4, 0.7_8_1_2_2_2_6, 0.7_6_9_4_2_5_1_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Any = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Any = EulerDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Union[str, Any] = self.get_dummy_inputs() _lowercase : List[str] = pipe(**lowerCamelCase).images _lowercase : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Any = np.array( [0.6_9_7_4_7_8_2, 0.6_8_9_0_2_0_9_3, 0.7_0_1_3_5_8_8_5, 0.7_5_8_3_6_1_8, 0.7_8_0_4_5_4_5, 0.7_8_5_4_9_1_2, 0.7_8_6_6_7_4_2_6, 0.7_8_7_4_3_8_6_3, 0.7_8_0_7_0_2_2_3]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : List[Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint, provider='CPUExecutionProvider') _lowercase : Union[str, Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[str] = self.get_dummy_inputs() _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images _lowercase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _lowercase : Dict = np.array( [0.7_7_4_2_4_4_9_6, 0.7_7_3_6_0_1, 0.7_6_4_5_2_8_8, 0.7_7_6_9_5_9_8, 0.7_7_7_2_7_3_9, 0.7_7_3_8_6_8_8, 0.7_8_1_8_7_2_3_3, 0.7_7_8_7_9_5_8_4, 0.7_6_7_0_4_3]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): @property def UpperCamelCase ( self) -> str: """simple docstring""" return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def UpperCamelCase ( self) -> Optional[Any]: """simple docstring""" _lowercase : List[str] = ort.SessionOptions() _lowercase : Union[str, Any] = False return options def UpperCamelCase ( self) -> List[str]: """simple docstring""" _lowercase : Optional[int] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg') _lowercase : List[Any] = init_image.resize((1_28, 1_28)) # using the PNDM scheduler by default _lowercase : int = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx', provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = 'A fantasy landscape, trending on artstation' _lowercase : Any = torch.manual_seed(0) _lowercase : List[Any] = pipe( prompt=lowerCamelCase, image=lowerCamelCase, guidance_scale=7.5, num_inference_steps=10, generator=lowerCamelCase, output_type='np', ) _lowercase : Tuple = output.images _lowercase : Optional[Any] = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 5_12, 3) _lowercase : Any = np.array([0.4_8_8_3, 0.4_9_4_7, 0.4_9_8_0, 0.4_9_7_5, 0.4_9_8_2, 0.4_9_8_0, 0.5_0_0_0, 0.5_0_0_6, 0.4_9_7_2]) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice).max() < 2E-2 def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Union[str, Any] = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/img2img/sketch-mountains-input.jpg') _lowercase : List[Any] = init_image.resize((1_28, 1_28)) _lowercase : Optional[Any] = LMSDiscreteScheduler.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx', subfolder='scheduler') _lowercase : int = OnnxStableDiffusionUpscalePipeline.from_pretrained( 'ssube/stable-diffusion-x4-upscaler-onnx', scheduler=lowerCamelCase, provider=self.gpu_provider, sess_options=self.gpu_options, ) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Union[str, Any] = 'A fantasy landscape, trending on artstation' _lowercase : int = torch.manual_seed(0) _lowercase : Union[str, Any] = pipe( prompt=lowerCamelCase, image=lowerCamelCase, guidance_scale=7.5, num_inference_steps=20, generator=lowerCamelCase, output_type='np', ) _lowercase : Tuple = output.images _lowercase : Tuple = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 5_12, 3) _lowercase : Optional[int] = np.array( [0.5_0_1_7_3_7_5_3, 0.5_0_2_2_3_3_5_6, 0.5_0_2_0_3_9, 0.5_0_2_3_3_0_3_6, 0.5_0_2_3_7_2_5, 0.5_0_2_2_6_0_1, 0.5_0_1_8_7_5_8, 0.5_0_2_3_4_0_8_5, 0.5_0_2_4_1_5_6_6]) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice).max() < 2E-2
84
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE : List[Any] = {"configuration_deit": ["DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DeiTConfig", "DeiTOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : Dict = ["DeiTFeatureExtractor"] SCREAMING_SNAKE_CASE : Dict = ["DeiTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : str = [ "DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "DeiTForImageClassification", "DeiTForImageClassificationWithTeacher", "DeiTForMaskedImageModeling", "DeiTModel", "DeiTPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE : List[str] = [ "TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFDeiTForImageClassification", "TFDeiTForImageClassificationWithTeacher", "TFDeiTForMaskedImageModeling", "TFDeiTModel", "TFDeiTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_deit import DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP, DeiTConfig, DeiTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_deit import DeiTFeatureExtractor from .image_processing_deit import DeiTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deit import ( DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, DeiTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deit import ( TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, TFDeiTPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
84
1
def lowerCAmelCase_ ( _snake_case : Tuple , _snake_case : List[str] ) -> List[str]: '''simple docstring''' __magic_name__ : Union[str, Any] = '''''' for i in table: res += inp[i - 1] return res def lowerCAmelCase_ ( _snake_case : Tuple ) -> Union[str, Any]: '''simple docstring''' return data[1:] + data[0] def lowerCAmelCase_ ( _snake_case : List[str] , _snake_case : Tuple ) -> Dict: '''simple docstring''' __magic_name__ : Any = '''''' for i in range(len(_snake_case ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def lowerCAmelCase_ ( _snake_case : int , _snake_case : int ) -> Dict: '''simple docstring''' __magic_name__ : List[str] = int("0b" + data[0] + data[-1] , 2 ) __magic_name__ : str = int("0b" + data[1:3] , 2 ) return bin(s[row][col] )[2:] def lowerCAmelCase_ ( _snake_case : int , _snake_case : Optional[int] , _snake_case : Optional[int] , _snake_case : Union[str, Any] , _snake_case : Optional[int] ) -> Any: '''simple docstring''' __magic_name__ : Dict = message[:4] __magic_name__ : List[Any] = message[4:] __magic_name__ : Tuple = apply_table(_snake_case , _snake_case ) __magic_name__ : Tuple = xor(_snake_case , _snake_case ) __magic_name__ : str = apply_sbox(_snake_case , temp[:4] ) # noqa: E741 __magic_name__ : List[str] = apply_sbox(_snake_case , temp[4:] ) __magic_name__ : Tuple = '''0''' * (2 - len(_snake_case )) + l # noqa: E741 __magic_name__ : Union[str, Any] = '''0''' * (2 - len(_snake_case )) + r __magic_name__ : str = apply_table(l + r , _snake_case ) __magic_name__ : int = xor(_snake_case , _snake_case ) return temp + right if __name__ == "__main__": snake_case : Optional[int] = input("Enter 10 bit key: ") snake_case : Union[str, Any] = input("Enter 8 bit message: ") snake_case : Any = [6, 3, 7, 4, 8, 5, 10, 9] snake_case : Dict = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] snake_case : Union[str, Any] = [2, 4, 3, 1] snake_case : Union[str, Any] = [2, 6, 3, 1, 4, 8, 5, 7] snake_case : List[Any] = [4, 1, 3, 5, 7, 2, 8, 6] snake_case : Optional[int] = [4, 1, 2, 3, 2, 3, 4, 1] snake_case : int = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] snake_case : Optional[int] = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation snake_case : Tuple = apply_table(key, paa_table) snake_case : Any = temp[:5] snake_case : Any = temp[5:] snake_case : Optional[Any] = left_shift(left) snake_case : int = left_shift(right) snake_case : Tuple = apply_table(left + right, pa_table) snake_case : Union[str, Any] = left_shift(left) snake_case : str = left_shift(right) snake_case : Dict = left_shift(left) snake_case : Any = left_shift(right) snake_case : int = apply_table(left + right, pa_table) # encryption snake_case : List[Any] = apply_table(message, IP) snake_case : List[Any] = function(expansion, sa, sa, keya, temp) snake_case : Dict = temp[4:] + temp[:4] snake_case : List[str] = function(expansion, sa, sa, keya, temp) snake_case : Any = apply_table(temp, IP_inv) print("Cipher text is:", CT) # decryption snake_case : Union[str, Any] = apply_table(CT, IP) snake_case : Any = function(expansion, sa, sa, keya, temp) snake_case : List[Any] = temp[4:] + temp[:4] snake_case : Optional[int] = function(expansion, sa, sa, keya, temp) snake_case : Any = apply_table(temp, IP_inv) print("Plain text after decypting is:", PT)
281
'''simple docstring''' from __future__ import annotations import math def __UpperCAmelCase ( A : int , A : int , A : bool , A : list[int] , A : float ) -> int: if depth < 0: raise ValueError('''Depth cannot be less than 0''' ) if not scores: raise ValueError('''Scores cannot be empty''' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , A , A , A ) , minimax(depth + 1 , node_index * 2 + 1 , A , A , A ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , A , A , A ) , minimax(depth + 1 , node_index * 2 + 1 , A , A , A ) , ) ) def __UpperCAmelCase ( ) -> None: UpperCAmelCase_ : List[str] = [9_0, 2_3, 6, 3_3, 2_1, 6_5, 1_2_3, 3_4_4_2_3] UpperCAmelCase_ : List[Any] = math.log(len(A ) , 2 ) print(F"Optimal value : {minimax(0 , 0 , A , A , A )}" ) if __name__ == "__main__": import doctest doctest.testmod() main()
304
0
from ..utils import DummyObject, requires_backends class a__ ( metaclass=__snake_case ): A__ : int = ['note_seq'] def __init__( self , *UpperCAmelCase , **UpperCAmelCase ) -> List[str]: requires_backends(self , ['note_seq'] ) @classmethod def __SCREAMING_SNAKE_CASE ( cls , *UpperCAmelCase , **UpperCAmelCase ) -> Optional[int]: requires_backends(cls , ['note_seq'] ) @classmethod def __SCREAMING_SNAKE_CASE ( cls , *UpperCAmelCase , **UpperCAmelCase ) -> Tuple: requires_backends(cls , ['note_seq'] )
197
def lowerCAmelCase( __lowerCamelCase ): if not all(char in '01' for char in bin_string ): raise ValueError('Non-binary value was passed to the function' ) if not bin_string: raise ValueError('Empty string was passed to the function' ) __a = '' while len(__lowerCamelCase ) % 3 != 0: __a = '0' + bin_string __a = [ bin_string[index : index + 3] for index in range(len(__lowerCamelCase ) ) if index % 3 == 0 ] for bin_group in bin_string_in_3_list: __a = 0 for index, val in enumerate(__lowerCamelCase ): oct_val += int(2 ** (2 - index) * int(__lowerCamelCase ) ) oct_string += str(__lowerCamelCase ) return oct_string if __name__ == "__main__": from doctest import testmod testmod()
197
1
'''simple docstring''' from __future__ import annotations from decimal import Decimal from numpy import array def a ( __a ) -> list[list[float]]: '''simple docstring''' UpperCamelCase__ :List[str] = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(__a ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix UpperCamelCase__ :Optional[int] = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError('''This matrix has no inverse.''' ) # Creates a copy of the matrix with swapped positions of the elements UpperCamelCase__ :List[Any] = [[0.0, 0.0], [0.0, 0.0]] UpperCamelCase__ , UpperCamelCase__ :int = matrix[1][1], matrix[0][0] UpperCamelCase__ , UpperCamelCase__ :Union[str, Any] = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(__a ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(__a ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule UpperCamelCase__ :Tuple = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError('''This matrix has no inverse.''' ) # Creating cofactor matrix UpperCamelCase__ :Any = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] UpperCamelCase__ :int = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) UpperCamelCase__ :Union[str, Any] = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) UpperCamelCase__ :Tuple = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) UpperCamelCase__ :Any = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) UpperCamelCase__ :Dict = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) UpperCamelCase__ :Tuple = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) UpperCamelCase__ :List[Any] = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) UpperCamelCase__ :str = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) UpperCamelCase__ :Tuple = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) UpperCamelCase__ :Optional[int] = array(__a ) for i in range(3 ): for j in range(3 ): UpperCamelCase__ :Optional[int] = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix UpperCamelCase__ :str = array(__a ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(__a ) # Calculate the inverse of the matrix return [[float(d(__a ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError('''Please provide a matrix of size 2x2 or 3x3.''' )
97
import csv import tweepy # Twitter API credentials a ="""""" a ="""""" a ="""""" a ="""""" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> None: # authorize twitter, initialize tweepy __lowerCamelCase : Tuple = tweepy.OAuthHandler(lowerCamelCase__ , lowerCamelCase__ ) auth.set_access_token(lowerCamelCase__ , lowerCamelCase__ ) __lowerCamelCase : Optional[int] = tweepy.API(lowerCamelCase__ ) # initialize a list to hold all the tweepy Tweets __lowerCamelCase : str = [] # make initial request for most recent tweets (200 is the maximum allowed count) __lowerCamelCase : Union[str, Any] = api.user_timeline(screen_name=lowerCamelCase__ , count=2_0_0 ) # save most recent tweets alltweets.extend(lowerCamelCase__ ) # save the id of the oldest tweet less one __lowerCamelCase : Any = alltweets[-1].id - 1 # keep grabbing tweets until there are no tweets left to grab while len(lowerCamelCase__ ) > 0: print(F"getting tweets before {oldest}" ) # all subsequent requests use the max_id param to prevent duplicates __lowerCamelCase : str = api.user_timeline( screen_name=lowerCamelCase__ , count=2_0_0 , max_id=lowerCamelCase__ ) # save most recent tweets alltweets.extend(lowerCamelCase__ ) # update the id of the oldest tweet less one __lowerCamelCase : Optional[int] = alltweets[-1].id - 1 print(F"...{len(lowerCamelCase__ )} tweets downloaded so far" ) # transform the tweepy tweets into a 2D array that will populate the csv __lowerCamelCase : str = [[tweet.id_str, tweet.created_at, tweet.text] for tweet in alltweets] # write the csv with open(F"new_{screen_name}_tweets.csv" , 'w' ) as f: __lowerCamelCase : Any = csv.writer(lowerCamelCase__ ) writer.writerow(['id', 'created_at', 'text'] ) writer.writerows(lowerCamelCase__ ) if __name__ == "__main__": # pass in the username of the account you want to download get_all_tweets("""FirePing32""")
73
0
'''simple docstring''' import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('1.6'): _snake_case = True from torch.cuda.amp import autocast _snake_case = logging.getLogger(__name__) def _A ( snake_case=None , snake_case=None ) -> Any: return field(default_factory=lambda: default , metadata=snake_case ) @dataclass class a__ : _SCREAMING_SNAKE_CASE : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) _SCREAMING_SNAKE_CASE : Optional[str] = field( default=lowerCamelCase_ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) _SCREAMING_SNAKE_CASE : Optional[bool] = field( default=lowerCamelCase_ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) _SCREAMING_SNAKE_CASE : Optional[float] = field( default=0.1 , metadata={'help': 'The dropout ratio for the attention probabilities.'} ) _SCREAMING_SNAKE_CASE : Optional[float] = field( default=0.1 , metadata={'help': 'The dropout ratio for activations inside the fully connected layer.'} ) _SCREAMING_SNAKE_CASE : Optional[float] = field( default=0.1 , metadata={ 'help': 'The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.' } , ) _SCREAMING_SNAKE_CASE : Optional[float] = field( default=0.1 , metadata={'help': 'The dropout probabilitiy for all 1D convolutional layers in feature extractor.'} , ) _SCREAMING_SNAKE_CASE : Optional[float] = field( default=0.05 , metadata={ 'help': ( 'Propability of each feature vector along the time axis to be chosen as the start of the vector' 'span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature' 'vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.' ) } , ) _SCREAMING_SNAKE_CASE : Optional[float] = field(default=0.0 , metadata={'help': 'The LayerDrop probability.'} ) @dataclass class a__ : _SCREAMING_SNAKE_CASE : Optional[str] = field( default=lowerCamelCase_ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) _SCREAMING_SNAKE_CASE : Optional[str] = field( default='train+validation' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) _SCREAMING_SNAKE_CASE : bool = field( default=lowerCamelCase_ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) _SCREAMING_SNAKE_CASE : Optional[int] = field( default=lowerCamelCase_ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) _SCREAMING_SNAKE_CASE : Optional[int] = field( default=lowerCamelCase_ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) _SCREAMING_SNAKE_CASE : Optional[int] = field( default=lowerCamelCase_ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of validation examples to this ' 'value if set.' ) } , ) _SCREAMING_SNAKE_CASE : List[str] = list_field( default=[',', '?', '.', '!', '-', ';', ':', '""', '%', '\'', '"', '�'] , metadata={'help': 'A list of characters to remove from the transcripts.'} , ) @dataclass class a__ : _SCREAMING_SNAKE_CASE : WavaVecaProcessor _SCREAMING_SNAKE_CASE : Union[bool, str] = True _SCREAMING_SNAKE_CASE : Optional[int] = None _SCREAMING_SNAKE_CASE : Optional[int] = None _SCREAMING_SNAKE_CASE : Optional[int] = None _SCREAMING_SNAKE_CASE : Optional[int] = None def __call__( self , _UpperCamelCase ): """simple docstring""" _lowercase : List[str] = [{"input_values": feature["input_values"]} for feature in features] _lowercase : Dict = [{"input_ids": feature["labels"]} for feature in features] _lowercase : Any = self.processor.pad( _UpperCamelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="pt" , ) _lowercase : Union[str, Any] = self.processor.pad( labels=_UpperCamelCase , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors="pt" , ) # replace padding with -100 to ignore loss correctly _lowercase : List[str] = labels_batch["input_ids"].masked_fill(labels_batch.attention_mask.ne(1 ) , -100 ) _lowercase : Optional[Any] = labels return batch class a__ ( lowerCamelCase_ ): def _lowerCamelCase ( self , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" model.train() _lowercase : str = self._prepare_inputs(_UpperCamelCase ) if self.use_amp: with autocast(): _lowercase : Optional[int] = self.compute_loss(_UpperCamelCase , _UpperCamelCase ) else: _lowercase : Tuple = self.compute_loss(_UpperCamelCase , _UpperCamelCase ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": _lowercase : int = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": _lowercase : Any = loss.sum() / (inputs["labels"] >= 0).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: _lowercase : Optional[int] = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(_UpperCamelCase ).backward() elif self.use_apex: with amp.scale_loss(_UpperCamelCase , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(_UpperCamelCase ) else: loss.backward() return loss.detach() def _A ( ) -> Optional[Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _lowercase : Optional[int] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowercase : List[Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowercase : int = parser.parse_args_into_dataclasses() # Detecting last checkpoint. _lowercase : Union[str, Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowercase : Union[str, Any] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info("Training/evaluation parameters %s" , snake_case ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: _lowercase : Tuple = datasets.load_dataset( "common_voice" , data_args.dataset_config_name , split=data_args.train_split_name ) _lowercase : Any = datasets.load_dataset("common_voice" , data_args.dataset_config_name , split="test" ) # Create and save tokenizer _lowercase : Dict = F'''[{''.join(data_args.chars_to_ignore )}]''' def remove_special_characters(snake_case ): _lowercase : List[str] = re.sub(snake_case , "" , batch["sentence"] ).lower() + " " return batch _lowercase : int = train_dataset.map(snake_case , remove_columns=["sentence"] ) _lowercase : int = eval_dataset.map(snake_case , remove_columns=["sentence"] ) def extract_all_chars(snake_case ): _lowercase : Optional[int] = " ".join(batch["text"] ) _lowercase : int = list(set(snake_case ) ) return {"vocab": [vocab], "all_text": [all_text]} _lowercase : Dict = train_dataset.map( snake_case , batched=snake_case , batch_size=-1 , keep_in_memory=snake_case , remove_columns=train_dataset.column_names , ) _lowercase : List[Any] = train_dataset.map( snake_case , batched=snake_case , batch_size=-1 , keep_in_memory=snake_case , remove_columns=eval_dataset.column_names , ) _lowercase : Dict = list(set(vocab_train["vocab"][0] ) | set(vocab_test["vocab"][0] ) ) _lowercase : List[str] = {v: k for k, v in enumerate(snake_case )} _lowercase : Union[str, Any] = vocab_dict[" "] del vocab_dict[" "] _lowercase : Dict = len(snake_case ) _lowercase : Dict = len(snake_case ) with open("vocab.json" , "w" ) as vocab_file: json.dump(snake_case , snake_case ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowercase : Dict = WavaVecaCTCTokenizer( "vocab.json" , unk_token="[UNK]" , pad_token="[PAD]" , word_delimiter_token="|" , ) _lowercase : List[str] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0.0 , do_normalize=snake_case , return_attention_mask=snake_case ) _lowercase : Optional[int] = WavaVecaProcessor(feature_extractor=snake_case , tokenizer=snake_case ) _lowercase : int = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction="mean" , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: _lowercase : Optional[Any] = min(len(snake_case ) , data_args.max_train_samples ) _lowercase : Any = train_dataset.select(range(snake_case ) ) if data_args.max_val_samples is not None: _lowercase : Any = eval_dataset.select(range(data_args.max_val_samples ) ) _lowercase : Dict = torchaudio.transforms.Resample(4_80_00 , 1_60_00 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(snake_case ): _lowercase : List[str] = torchaudio.load(batch["path"] ) _lowercase : List[Any] = resampler(snake_case ).squeeze().numpy() _lowercase : str = 1_60_00 _lowercase : Optional[int] = batch["text"] return batch _lowercase : Dict = train_dataset.map( snake_case , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) _lowercase : List[str] = eval_dataset.map( snake_case , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(snake_case ): # check that all files have the correct sampling rate assert ( len(set(batch["sampling_rate"] ) ) == 1 ), F'''Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.''' _lowercase : str = processor( audio=batch["speech"] , text=batch["target_text"] , sampling_rate=batch["sampling_rate"][0] ) batch.update(snake_case ) return batch _lowercase : Any = train_dataset.map( snake_case , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=snake_case , num_proc=data_args.preprocessing_num_workers , ) _lowercase : Dict = eval_dataset.map( snake_case , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=snake_case , num_proc=data_args.preprocessing_num_workers , ) # Metric _lowercase : Optional[int] = datasets.load_metric("wer" ) def compute_metrics(snake_case ): _lowercase : Dict = pred.predictions _lowercase : int = np.argmax(snake_case , axis=-1 ) _lowercase : str = processor.tokenizer.pad_token_id _lowercase : Optional[int] = processor.batch_decode(snake_case ) # we do not want to group tokens when computing the metrics _lowercase : int = processor.batch_decode(pred.label_ids , group_tokens=snake_case ) _lowercase : Optional[int] = wer_metric.compute(predictions=snake_case , references=snake_case ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator _lowercase : str = DataCollatorCTCWithPadding(processor=snake_case , padding=snake_case ) # Initialize our Trainer _lowercase : List[str] = CTCTrainer( model=snake_case , data_collator=snake_case , args=snake_case , compute_metrics=snake_case , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: _lowercase : Any = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): _lowercase : Tuple = model_args.model_name_or_path else: _lowercase : Any = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) _lowercase : str = trainer.train(resume_from_checkpoint=snake_case ) trainer.save_model() _lowercase : List[str] = train_result.metrics _lowercase : List[str] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(snake_case ) ) _lowercase : Any = min(snake_case , len(snake_case ) ) trainer.log_metrics("train" , snake_case ) trainer.save_metrics("train" , snake_case ) trainer.save_state() # Evaluation _lowercase : Dict = {} if training_args.do_eval: logger.info("*** Evaluate ***" ) _lowercase : Optional[Any] = trainer.evaluate() _lowercase : Optional[Any] = data_args.max_val_samples if data_args.max_val_samples is not None else len(snake_case ) _lowercase : Tuple = min(snake_case , len(snake_case ) ) trainer.log_metrics("eval" , snake_case ) trainer.save_metrics("eval" , snake_case ) return results if __name__ == "__main__": main()
366
'''simple docstring''' import os import sys import unittest _snake_case = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path _snake_case = os.path.join(git_repo_path, 'src', 'transformers') _snake_case = '\n{0} = None\n' _snake_case = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n' _snake_case = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' class a__ ( unittest.TestCase ): def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Optional[int] = find_backend(" _import_structure[\"models.albert\"].append(\"AlbertTokenizerFast\")" ) self.assertIsNone(_UpperCamelCase ) _lowercase : Dict = find_backend(" if not is_tokenizers_available():" ) self.assertEqual(_UpperCamelCase , "tokenizers" ) _lowercase : str = find_backend(" if not is_tensorflow_text_available():" ) self.assertEqual(_UpperCamelCase , "tensorflow_text" ) _lowercase : Optional[Any] = find_backend(" if not (is_sentencepiece_available() and is_tokenizers_available()):" ) self.assertEqual(_UpperCamelCase , "sentencepiece_and_tokenizers" ) _lowercase : List[str] = find_backend( " if not (is_sentencepiece_available() and is_tensorflow_text_available()):" ) self.assertEqual(_UpperCamelCase , "sentencepiece_and_tensorflow_text" ) _lowercase : Optional[Any] = find_backend( " if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):" ) self.assertEqual(_UpperCamelCase , "sentencepiece_and_tokenizers_and_vision" ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : List[Any] = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn("torch" , _UpperCamelCase ) self.assertIn("tensorflow_text" , _UpperCamelCase ) self.assertIn("sentencepiece_and_tokenizers" , _UpperCamelCase ) # Likewise, we can't assert on the exact content of a key self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertModel" , objects["tf"] ) self.assertIn("FlaxBertModel" , objects["flax"] ) self.assertIn("BertModel" , objects["torch"] ) self.assertIn("TFBertTokenizer" , objects["tensorflow_text"] ) self.assertIn("convert_slow_tokenizer" , objects["sentencepiece_and_tokenizers"] ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : int = create_dummy_object("CONSTANT" , "'torch'" ) self.assertEqual(_UpperCamelCase , "\nCONSTANT = None\n" ) _lowercase : Optional[Any] = create_dummy_object("function" , "'torch'" ) self.assertEqual( _UpperCamelCase , "\ndef function(*args, **kwargs):\n requires_backends(function, 'torch')\n" ) _lowercase : Union[str, Any] = "\nclass FakeClass(metaclass=DummyObject):\n _backends = 'torch'\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, 'torch')\n" _lowercase : List[str] = create_dummy_object("FakeClass" , "'torch'" ) self.assertEqual(_UpperCamelCase , _UpperCamelCase ) def _lowerCamelCase ( self ): """simple docstring""" _lowercase : Optional[int] = "# This file is autogenerated by the command `make fix-copies`, do not edit.\nfrom ..utils import DummyObject, requires_backends\n\n\nCONSTANT = None\n\n\ndef function(*args, **kwargs):\n requires_backends(function, [\"torch\"])\n\n\nclass FakeClass(metaclass=DummyObject):\n _backends = [\"torch\"]\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, [\"torch\"])\n" _lowercase : str = create_dummy_files({"torch": ["CONSTANT", "function", "FakeClass"]} ) self.assertEqual(dummy_files["torch"] , _UpperCamelCase )
199
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_fnet import FNetTokenizer else: A_ :Optional[int] = None A_ :int = logging.get_logger(__name__) A_ :Any = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} A_ :Dict = { '''vocab_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/spiece.model''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''google/fnet-base''': '''https://huggingface.co/google/fnet-base/resolve/main/tokenizer.json''', '''google/fnet-large''': '''https://huggingface.co/google/fnet-large/resolve/main/tokenizer.json''', }, } A_ :Optional[int] = { '''google/fnet-base''': 512, '''google/fnet-large''': 512, } A_ :Optional[int] = '''▁''' class __A ( a ): """simple docstring""" UpperCamelCase__ : Union[str, Any] =VOCAB_FILES_NAMES UpperCamelCase__ : Union[str, Any] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : List[Any] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : Tuple =["""input_ids""", """token_type_ids"""] UpperCamelCase__ : Tuple =FNetTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="<unk>" , lowerCamelCase__="[SEP]" , lowerCamelCase__="<pad>" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , **lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[Any] =( AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ , normalized=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token ) super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , remove_space=lowerCamelCase__ , keep_accents=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , **lowerCamelCase__ , ) __UpperCamelCase : List[str] =do_lower_case __UpperCamelCase : Tuple =remove_space __UpperCamelCase : List[str] =keep_accents __UpperCamelCase : int =vocab_file __UpperCamelCase : Dict =False if not self.vocab_file else True def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : Optional[Any] =[self.sep_token_id] __UpperCamelCase : Any =[self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : Union[str, Any] =[self.sep_token_id] __UpperCamelCase : Tuple =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" if not os.path.isdir(lowerCamelCase__ ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return __UpperCamelCase : Optional[int] =os.path.join( lowerCamelCase__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase__ ): copyfile(self.vocab_file , lowerCamelCase__ ) return (out_vocab_file,)
71
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class A (unittest.TestCase ): '''simple docstring''' def __init__( self : List[str] , __lowerCAmelCase : int , __lowerCAmelCase : List[str]=13 , __lowerCAmelCase : int=7 , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : int=True , __lowerCAmelCase : Dict=True , __lowerCAmelCase : Union[str, Any]=True , __lowerCAmelCase : List[Any]=99 , __lowerCAmelCase : Optional[Any]=32 , __lowerCAmelCase : Optional[Any]=5 , __lowerCAmelCase : Tuple=4 , __lowerCAmelCase : Any=37 , __lowerCAmelCase : str="gelu" , __lowerCAmelCase : Optional[Any]=0.1 , __lowerCAmelCase : Union[str, Any]=0.1 , __lowerCAmelCase : List[str]=5_12 , __lowerCAmelCase : int=16 , __lowerCAmelCase : Optional[int]=2 , __lowerCAmelCase : List[Any]=0.0_2 , __lowerCAmelCase : Tuple=4 , ) -> Dict: """simple docstring""" A__ = parent A__ = batch_size A__ = seq_length A__ = is_training A__ = use_attention_mask A__ = use_token_type_ids A__ = use_labels A__ = vocab_size A__ = hidden_size A__ = num_hidden_layers A__ = num_attention_heads A__ = intermediate_size A__ = hidden_act A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = max_position_embeddings A__ = type_vocab_size A__ = type_sequence_label_size A__ = initializer_range A__ = num_choices def a_ ( self : Any ) -> str: """simple docstring""" A__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ = None if self.use_attention_mask: A__ = random_attention_mask([self.batch_size, self.seq_length] ) A__ = None if self.use_token_type_ids: A__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def a_ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" A__ = self.prepare_config_and_inputs() A__ , A__ , A__ , A__ = config_and_inputs A__ = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict @require_flax class A (SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' __lowerCamelCase : str = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def a_ ( self : str ) -> Optional[int]: """simple docstring""" A__ = FlaxAlbertModelTester(self ) @slow def a_ ( self : int ) -> Tuple: """simple docstring""" for model_class_name in self.all_model_classes: A__ = model_class_name.from_pretrained("""albert-base-v2""" ) A__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowerCAmelCase ) @require_flax class A (unittest.TestCase ): '''simple docstring''' @slow def a_ ( self : Dict ) -> List[Any]: """simple docstring""" A__ = FlaxAlbertModel.from_pretrained("""albert-base-v2""" ) A__ = np.array([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) A__ = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) A__ = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase )[0] A__ = (1, 11, 7_68) self.assertEqual(output.shape , __lowerCAmelCase ) A__ = np.array( [[[-0.6_5_1_3, 1.5_0_3_5, -0.2_7_6_6], [-0.6_5_1_5, 1.5_0_4_6, -0.2_7_8_0], [-0.6_5_1_2, 1.5_0_4_9, -0.2_7_8_4]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , __lowerCAmelCase , atol=1e-4 ) )
274
0
def __A ( _lowercase ): '''simple docstring''' _A = [[0 for _ in range(_lowercase )] for _ in range(m + 1 )] for i in range(m + 1 ): _A = 1 for n in range(m + 1 ): for k in range(1 , _lowercase ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: __A = int(input('Enter a number: ').strip()) print(partition(n)) except ValueError: print('Please enter a number.') else: try: __A = int(sys.argv[1]) print(partition(n)) except ValueError: print('Please pass a number.')
361
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class SCREAMING_SNAKE_CASE ( snake_case , unittest.TestCase ): """simple docstring""" A_ = UnCLIPImageVariationPipeline A_ = IMAGE_VARIATION_PARAMS - {"height", "width", "guidance_scale"} A_ = IMAGE_VARIATION_BATCH_PARAMS A_ = [ "generator", "return_dict", "decoder_num_inference_steps", "super_res_num_inference_steps", ] A_ = False @property def __A ( self: Optional[Any] ) -> Optional[Any]: return 32 @property def __A ( self: List[str] ) -> Dict: return 32 @property def __A ( self: List[str] ) -> List[str]: return self.time_input_dim @property def __A ( self: Union[str, Any] ) -> Optional[int]: return self.time_input_dim * 4 @property def __A ( self: List[Any] ) -> Any: return 1_00 @property def __A ( self: List[str] ) -> Union[str, Any]: _A = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) return tokenizer @property def __A ( self: Optional[Any] ) -> Optional[Any]: torch.manual_seed(0 ) _A = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModelWithProjection(__A ) @property def __A ( self: List[str] ) -> int: torch.manual_seed(0 ) _A = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(__A ) @property def __A ( self: str ) -> List[str]: torch.manual_seed(0 ) _A = { '''clip_embeddings_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''cross_attention_dim''': self.cross_attention_dim, } _A = UnCLIPTextProjModel(**__A ) return model @property def __A ( self: Tuple ) -> str: torch.manual_seed(0 ) _A = { '''sample_size''': 32, # RGB in channels '''in_channels''': 3, # Out channels is double in channels because predicts mean and variance '''out_channels''': 6, '''down_block_types''': ('''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D'''), '''up_block_types''': ('''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''), '''mid_block_type''': '''UNetMidBlock2DSimpleCrossAttn''', '''block_out_channels''': (self.block_out_channels_a, self.block_out_channels_a * 2), '''layers_per_block''': 1, '''cross_attention_dim''': self.cross_attention_dim, '''attention_head_dim''': 4, '''resnet_time_scale_shift''': '''scale_shift''', '''class_embed_type''': '''identity''', } _A = UNetaDConditionModel(**__A ) return model @property def __A ( self: Tuple ) -> Any: return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def __A ( self: List[Any] ) -> Any: torch.manual_seed(0 ) _A = UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def __A ( self: List[Any] ) -> Dict: # seeded differently to get different unet than `self.dummy_super_res_first` torch.manual_seed(1 ) _A = UNetaDModel(**self.dummy_super_res_kwargs ) return model def __A ( self: List[str] ) -> str: _A = self.dummy_decoder _A = self.dummy_text_proj _A = self.dummy_text_encoder _A = self.dummy_tokenizer _A = self.dummy_super_res_first _A = self.dummy_super_res_last _A = UnCLIPScheduler( variance_type='''learned_range''' , prediction_type='''epsilon''' , num_train_timesteps=10_00 , ) _A = UnCLIPScheduler( variance_type='''fixed_small_log''' , prediction_type='''epsilon''' , num_train_timesteps=10_00 , ) _A = CLIPImageProcessor(crop_size=32 , size=32 ) _A = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def __A ( self: Dict , __A: List[str] , __A: Any=0 , __A: Union[str, Any]=True ) -> Optional[Any]: _A = floats_tensor((1, 3, 32, 32) , rng=random.Random(__A ) ).to(__A ) if str(__A ).startswith('''mps''' ): _A = torch.manual_seed(__A ) else: _A = torch.Generator(device=__A ).manual_seed(__A ) if pil_image: _A = input_image * 0.5 + 0.5 _A = input_image.clamp(0 , 1 ) _A = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() _A = DiffusionPipeline.numpy_to_pil(__A )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def __A ( self: List[str] ) -> Union[str, Any]: _A = '''cpu''' _A = self.get_dummy_components() _A = self.pipeline_class(**__A ) _A = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) _A = self.get_dummy_inputs(__A , pil_image=__A ) _A = pipe(**__A ) _A = output.images _A = self.get_dummy_inputs(__A , pil_image=__A ) _A = pipe( **__A , return_dict=__A , )[0] _A = image[0, -3:, -3:, -1] _A = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _A = np.array( [ 0.9_997, 0.0_002, 0.9_997, 0.9_997, 0.9_969, 0.0_023, 0.9_997, 0.9_969, 0.9_970, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self: Optional[int] ) -> Tuple: _A = '''cpu''' _A = self.get_dummy_components() _A = self.pipeline_class(**__A ) _A = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) _A = self.get_dummy_inputs(__A , pil_image=__A ) _A = pipe(**__A ) _A = output.images _A = self.get_dummy_inputs(__A , pil_image=__A ) _A = pipe( **__A , return_dict=__A , )[0] _A = image[0, -3:, -3:, -1] _A = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _A = np.array([0.9_997, 0.0_003, 0.9_997, 0.9_997, 0.9_970, 0.0_024, 0.9_997, 0.9_971, 0.9_971] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self: Any ) -> Dict: _A = '''cpu''' _A = self.get_dummy_components() _A = self.pipeline_class(**__A ) _A = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) _A = self.get_dummy_inputs(__A , pil_image=__A ) _A = [ pipeline_inputs['''image'''], pipeline_inputs['''image'''], ] _A = pipe(**__A ) _A = output.images _A = self.get_dummy_inputs(__A , pil_image=__A ) _A = [ tuple_pipeline_inputs['''image'''], tuple_pipeline_inputs['''image'''], ] _A = pipe( **__A , return_dict=__A , )[0] _A = image[0, -3:, -3:, -1] _A = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) _A = np.array( [ 0.9_997, 0.9_989, 0.0_008, 0.0_021, 0.9_960, 0.0_018, 0.0_014, 0.0_002, 0.9_933, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 def __A ( self: List[str] ) -> Tuple: _A = torch.device('''cpu''' ) class SCREAMING_SNAKE_CASE : """simple docstring""" A_ = 1 _A = self.get_dummy_components() _A = self.pipeline_class(**__A ) _A = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) _A = torch.Generator(device=__A ).manual_seed(0 ) _A = pipe.decoder.dtype _A = 1 _A = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) _A = pipe.prepare_latents( __A , dtype=__A , device=__A , generator=__A , latents=__A , scheduler=DummyScheduler() ) _A = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) _A = pipe.prepare_latents( __A , dtype=__A , device=__A , generator=__A , latents=__A , scheduler=DummyScheduler() ) _A = self.get_dummy_inputs(__A , pil_image=__A ) _A = pipe( **__A , decoder_latents=__A , super_res_latents=__A ).images _A = self.get_dummy_inputs(__A , pil_image=__A ) # Don't pass image, instead pass embedding _A = pipeline_inputs.pop('''image''' ) _A = pipe.image_encoder(__A ).image_embeds _A = pipe( **__A , decoder_latents=__A , super_res_latents=__A , image_embeddings=__A , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1e-4 @skip_mps def __A ( self: Dict ) -> int: _A = torch_device == '''cpu''' # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor _A = 1e-2 self._test_attention_slicing_forward_pass( test_max_difference=__A , expected_max_diff=__A ) @skip_mps def __A ( self: Any ) -> str: _A = torch_device == '''cpu''' _A = True _A = [ '''decoder_num_inference_steps''', '''super_res_num_inference_steps''', ] self._test_inference_batch_single_identical( test_max_difference=__A , relax_max_difference=__A , additional_params_copy_to_batched_inputs=__A , ) def __A ( self: Dict ) -> Dict: _A = [ '''decoder_num_inference_steps''', '''super_res_num_inference_steps''', ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes _A = [2, 3] self._test_inference_batch_consistent( batch_sizes=__A , additional_params_copy_to_batched_inputs=__A , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=__A ) @skip_mps def __A ( self: Optional[int] ) -> Optional[Any]: return super().test_dict_tuple_outputs_equivalent() @skip_mps def __A ( self: Any ) -> Any: return super().test_save_load_local() @skip_mps def __A ( self: Tuple ) -> Union[str, Any]: return super().test_save_load_optional_components() @slow @require_torch_gpu class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __A ( self: Union[str, Any] ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __A ( self: int ) -> List[str]: _A = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png''' ) _A = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/unclip/karlo_v1_alpha_cat_variation_fp16.npy''' ) _A = UnCLIPImageVariationPipeline.from_pretrained( '''kakaobrain/karlo-v1-alpha-image-variations''' , torch_dtype=torch.floataa ) _A = pipeline.to(__A ) pipeline.set_progress_bar_config(disable=__A ) _A = torch.Generator(device='''cpu''' ).manual_seed(0 ) _A = pipeline( __A , generator=__A , output_type='''np''' , ) _A = output.images[0] assert image.shape == (2_56, 2_56, 3) assert_mean_pixel_difference(__A , __A , 15 )
75
0
def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = len(lowerCamelCase__ ) lowerCamelCase_ = [] for i in range(len(lowerCamelCase__ ) - pat_len + 1 ): lowerCamelCase_ = True for j in range(lowerCamelCase__ ): if s[i + j] != pattern[j]: lowerCamelCase_ = False break if match_found: position.append(lowerCamelCase__ ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
19
from collections import deque def lowerCamelCase_ ( lowerCamelCase__ ): lowerCamelCase_ = len(lowerCamelCase__ ) lowerCamelCase_ = deque() lowerCamelCase_ = [False for _ in range(lowerCamelCase__ )] lowerCamelCase_ = [-1 for _ in range(lowerCamelCase__ )] lowerCamelCase_ = index_of[:] def strong_connect(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = index # the number when this node is seen lowerCamelCase_ = index # lowest rank node reachable from here index += 1 stack.append(lowerCamelCase__ ) lowerCamelCase_ = True for w in g[v]: if index_of[w] == -1: lowerCamelCase_ = strong_connect(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) lowerCamelCase_ = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) elif on_stack[w]: lowerCamelCase_ = ( lowlink_of[w] if lowlink_of[w] < lowlink_of[v] else lowlink_of[v] ) if lowlink_of[v] == index_of[v]: lowerCamelCase_ = [] lowerCamelCase_ = stack.pop() lowerCamelCase_ = False component.append(lowerCamelCase__ ) while w != v: lowerCamelCase_ = stack.pop() lowerCamelCase_ = False component.append(lowerCamelCase__ ) components.append(lowerCamelCase__ ) return index lowerCamelCase_ = [] for v in range(lowerCamelCase__ ): if index_of[v] == -1: strong_connect(lowerCamelCase__ , 0 , lowerCamelCase__ ) return components def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = [[] for _ in range(lowerCamelCase__ )] for u, v in edges: g[u].append(lowerCamelCase__ ) return g if __name__ == "__main__": # Test __A =7 __A =[0, 0, 1, 2, 3, 3, 4, 4, 6] __A =[1, 3, 2, 0, 1, 4, 5, 6, 5] __A =[(u, v) for u, v in zip(source, target)] __A =create_graph(n_vertices, edges) assert [[5], [6], [4], [3, 2, 1, 0]] == tarjan(g)
19
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a : str= "▁" _a : int= {"vocab_file": "spiece.model"} _a : Tuple= { "vocab_file": {"google/pegasus-xsum": "https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model"} } _a : List[str]= { "google/pegasus-xsum": 512, } _a : int= logging.get_logger(__name__) class UpperCamelCase ( lowercase ): UpperCAmelCase : str = VOCAB_FILES_NAMES UpperCAmelCase : Optional[Any] = VOCAB_FILES_NAMES UpperCAmelCase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : int = ["""input_ids""", """attention_mask"""] def __init__(self : Tuple , _A : int , _A : List[str]="<pad>" , _A : List[str]="</s>" , _A : Optional[Any]="<unk>" , _A : Any="<mask_2>" , _A : Optional[Any]="<mask_1>" , _A : Union[str, Any]=None , _A : Optional[int]=1_03 , _A : Optional[Any] = None , **_A : List[str] , ) -> List[str]: __snake_case : str = offset if additional_special_tokens is not None: if not isinstance(_a , _a): raise TypeError( f"additional_special_tokens should be of type {type(_a)}, but is" f" {type(_a)}") __snake_case : str = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f"<unk_{i}>" for i in range(len(_a) , self.offset - 1) ] if len(set(_a)) != len(_a): raise ValueError( 'Please make sure that the provided additional_special_tokens do not contain an incorrectly' f" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.") __snake_case : Tuple = additional_special_tokens_extended else: __snake_case : Tuple = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f"<unk_{i}>" for i in range(2 , self.offset)] __snake_case : str = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=_a , unk_token=_a , mask_token=_a , pad_token=_a , mask_token_sent=_a , offset=_a , additional_special_tokens=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) __snake_case : Any = mask_token_sent __snake_case : List[str] = vocab_file __snake_case : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(_a) # add special tokens to encoder dict __snake_case : Dict[int, str] = { 0: self.pad_token, 1: self.eos_token, } if self.mask_token_sent is not None: self.encoder.update( { 2: self.mask_token_sent, 3: self.mask_token, }) if self.offset > 0: # entries 2-104 are only used for pretraining and called <mask_1>, <mask_2>, unk_2, ...unk_102 # mask_token_sent is already added to list -> so start at 1 self.encoder.update({i + 3: additional_special_tokens[i] for i in range(1 , self.offset - 1)}) __snake_case : Dict[str, int] = {v: k for k, v in self.encoder.items()} @property def _lowercase (self : Union[str, Any]) -> Any: return len(self.sp_model) + self.offset def _lowercase (self : Any) -> Optional[int]: __snake_case : Optional[int] = {self.convert_ids_to_tokens(_a): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__(self : List[Any]) -> Union[str, Any]: __snake_case : Dict = self.__dict__.copy() __snake_case : int = None return state def __setstate__(self : Union[str, Any] , _A : List[Any]) -> List[str]: __snake_case : List[Any] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs'): __snake_case : Optional[int] = {} __snake_case : List[str] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def _lowercase (self : Optional[int] , _A : int) -> Dict: return self.sp_model.encode(_a , out_type=_a) def _lowercase (self : Any , _A : str) -> Any: if token in self.decoder: return self.decoder[token] elif token in self.added_tokens_decoder: return self.added_tokens_decoder[token] __snake_case : Any = self.sp_model.piece_to_id(_a) return sp_id + self.offset def _lowercase (self : Dict , _A : Optional[int]) -> List[str]: if index in self.encoder: return self.encoder[index] elif index in self.added_tokens_encoder: return self.added_tokens_encoder[index] else: __snake_case : str = self.sp_model.IdToPiece(index - self.offset) return token def _lowercase (self : Any , _A : int) -> Dict: __snake_case : Union[str, Any] = [] __snake_case : int = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(_a) + token __snake_case : Optional[int] = [] else: current_sub_tokens.append(_a) out_string += self.sp_model.decode(_a) return out_string.strip() def _lowercase (self : str , _A : Any=False) -> List[Any]: return 1 def _lowercase (self : List[Any] , _A : Optional[Any]) -> int: __snake_case : Union[str, Any] = set(self.all_special_ids) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id) # <unk> is only sometimes special return [1 if x in all_special_ids else 0 for x in seq] def _lowercase (self : List[str] , _A : Any , _A : Union[str, Any] = None , _A : Optional[int] = False) -> Union[str, Any]: if already_has_special_tokens: return self._special_token_mask(_a) elif token_ids_a is None: return self._special_token_mask(_a) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a) + [1] def _lowercase (self : List[str] , _A : str , _A : str=None) -> List[str]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def _lowercase (self : Union[str, Any] , _A : int , _A : Optional[int] = None) -> Union[str, Any]: if not os.path.isdir(_a): logger.error(f"Vocabulary path ({save_directory}) should be a directory") return __snake_case : Any = os.path.join( _a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) if os.path.abspath(self.vocab_file) != os.path.abspath(_a) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , _a) elif not os.path.isfile(self.vocab_file): with open(_a , 'wb') as fi: __snake_case : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(_a) return (out_vocab_file,)
369
"""simple docstring""" import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor _a : Tuple= logging.get_logger(__name__) class UpperCamelCase ( lowercase ): def __init__(self : int , *_A : str , **_A : List[str]) -> None: warnings.warn( 'The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.' ' Please use VideoMAEImageProcessor instead.' , _A , ) super().__init__(*_A , **_A)
95
0
'''simple docstring''' import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _a : '''simple docstring''' @staticmethod def UpperCamelCase_ ( *A, **A ): '''simple docstring''' pass @is_pipeline_test @require_vision @require_torch class _a ( unittest.TestCase ): '''simple docstring''' A : Optional[Any] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def UpperCamelCase_ ( self, A, A, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = pipeline( 'zero-shot-object-detection', model='hf-internal-testing/tiny-random-owlvit-object-detection' ) SCREAMING_SNAKE_CASE : List[str] = [ { 'image': './tests/fixtures/tests_samples/COCO/000000039769.png', 'candidate_labels': ['cat', 'remote', 'couch'], } ] return object_detector, examples def UpperCamelCase_ ( self, A, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = object_detector(examples[0], threshold=0.0 ) SCREAMING_SNAKE_CASE : Tuple = len(A ) self.assertGreater(A, 0 ) self.assertEqual( A, [ { 'score': ANY(A ), 'label': ANY(A ), 'box': {'xmin': ANY(A ), 'ymin': ANY(A ), 'xmax': ANY(A ), 'ymax': ANY(A )}, } for i in range(A ) ], ) @require_tf @unittest.skip('Zero Shot Object Detection not implemented in TF' ) def UpperCamelCase_ ( self ): '''simple docstring''' pass @require_torch def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = pipeline( 'zero-shot-object-detection', model='hf-internal-testing/tiny-random-owlvit-object-detection' ) SCREAMING_SNAKE_CASE : str = object_detector( './tests/fixtures/tests_samples/COCO/000000039769.png', candidate_labels=['cat', 'remote', 'couch'], threshold=0.64, ) self.assertEqual( nested_simplify(A, decimals=4 ), [ {'score': 0.72_35, 'label': 'cat', 'box': {'xmin': 204, 'ymin': 167, 'xmax': 232, 'ymax': 190}}, {'score': 0.72_18, 'label': 'remote', 'box': {'xmin': 204, 'ymin': 167, 'xmax': 232, 'ymax': 190}}, {'score': 0.71_84, 'label': 'couch', 'box': {'xmin': 204, 'ymin': 167, 'xmax': 232, 'ymax': 190}}, {'score': 0.67_48, 'label': 'remote', 'box': {'xmin': 571, 'ymin': 83, 'xmax': 598, 'ymax': 103}}, {'score': 0.66_56, 'label': 'cat', 'box': {'xmin': 571, 'ymin': 83, 'xmax': 598, 'ymax': 103}}, {'score': 0.66_14, 'label': 'couch', 'box': {'xmin': 571, 'ymin': 83, 'xmax': 598, 'ymax': 103}}, {'score': 0.64_56, 'label': 'remote', 'box': {'xmin': 494, 'ymin': 105, 'xmax': 521, 'ymax': 127}}, {'score': 0.6_42, 'label': 'remote', 'box': {'xmin': 67, 'ymin': 274, 'xmax': 93, 'ymax': 297}}, {'score': 0.64_19, 'label': 'cat', 'box': {'xmin': 494, 'ymin': 105, 'xmax': 521, 'ymax': 127}}, ], ) SCREAMING_SNAKE_CASE : int = object_detector( [ { 'image': './tests/fixtures/tests_samples/COCO/000000039769.png', 'candidate_labels': ['cat', 'remote', 'couch'], } ], threshold=0.64, ) self.assertEqual( nested_simplify(A, decimals=4 ), [ [ {'score': 0.72_35, 'label': 'cat', 'box': {'xmin': 204, 'ymin': 167, 'xmax': 232, 'ymax': 190}}, {'score': 0.72_18, 'label': 'remote', 'box': {'xmin': 204, 'ymin': 167, 'xmax': 232, 'ymax': 190}}, {'score': 0.71_84, 'label': 'couch', 'box': {'xmin': 204, 'ymin': 167, 'xmax': 232, 'ymax': 190}}, {'score': 0.67_48, 'label': 'remote', 'box': {'xmin': 571, 'ymin': 83, 'xmax': 598, 'ymax': 103}}, {'score': 0.66_56, 'label': 'cat', 'box': {'xmin': 571, 'ymin': 83, 'xmax': 598, 'ymax': 103}}, {'score': 0.66_14, 'label': 'couch', 'box': {'xmin': 571, 'ymin': 83, 'xmax': 598, 'ymax': 103}}, {'score': 0.64_56, 'label': 'remote', 'box': {'xmin': 494, 'ymin': 105, 'xmax': 521, 'ymax': 127}}, {'score': 0.6_42, 'label': 'remote', 'box': {'xmin': 67, 'ymin': 274, 'xmax': 93, 'ymax': 297}}, {'score': 0.64_19, 'label': 'cat', 'box': {'xmin': 494, 'ymin': 105, 'xmax': 521, 'ymax': 127}}, ] ], ) @require_torch @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = pipeline('zero-shot-object-detection' ) SCREAMING_SNAKE_CASE : Optional[Any] = object_detector( 'http://images.cocodataset.org/val2017/000000039769.jpg', candidate_labels=['cat', 'remote', 'couch'], ) self.assertEqual( nested_simplify(A, decimals=4 ), [ {'score': 0.28_68, 'label': 'cat', 'box': {'xmin': 324, 'ymin': 20, 'xmax': 640, 'ymax': 373}}, {'score': 0.2_77, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 72, 'xmax': 177, 'ymax': 115}}, {'score': 0.25_37, 'label': 'cat', 'box': {'xmin': 1, 'ymin': 55, 'xmax': 315, 'ymax': 472}}, {'score': 0.14_74, 'label': 'remote', 'box': {'xmin': 335, 'ymin': 74, 'xmax': 371, 'ymax': 187}}, {'score': 0.12_08, 'label': 'couch', 'box': {'xmin': 4, 'ymin': 0, 'xmax': 642, 'ymax': 476}}, ], ) SCREAMING_SNAKE_CASE : List[Any] = object_detector( [ { 'image': 'http://images.cocodataset.org/val2017/000000039769.jpg', 'candidate_labels': ['cat', 'remote', 'couch'], }, { 'image': 'http://images.cocodataset.org/val2017/000000039769.jpg', 'candidate_labels': ['cat', 'remote', 'couch'], }, ], ) self.assertEqual( nested_simplify(A, decimals=4 ), [ [ {'score': 0.28_68, 'label': 'cat', 'box': {'xmin': 324, 'ymin': 20, 'xmax': 640, 'ymax': 373}}, {'score': 0.2_77, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 72, 'xmax': 177, 'ymax': 115}}, {'score': 0.25_37, 'label': 'cat', 'box': {'xmin': 1, 'ymin': 55, 'xmax': 315, 'ymax': 472}}, {'score': 0.14_74, 'label': 'remote', 'box': {'xmin': 335, 'ymin': 74, 'xmax': 371, 'ymax': 187}}, {'score': 0.12_08, 'label': 'couch', 'box': {'xmin': 4, 'ymin': 0, 'xmax': 642, 'ymax': 476}}, ], [ {'score': 0.28_68, 'label': 'cat', 'box': {'xmin': 324, 'ymin': 20, 'xmax': 640, 'ymax': 373}}, {'score': 0.2_77, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 72, 'xmax': 177, 'ymax': 115}}, {'score': 0.25_37, 'label': 'cat', 'box': {'xmin': 1, 'ymin': 55, 'xmax': 315, 'ymax': 472}}, {'score': 0.14_74, 'label': 'remote', 'box': {'xmin': 335, 'ymin': 74, 'xmax': 371, 'ymax': 187}}, {'score': 0.12_08, 'label': 'couch', 'box': {'xmin': 4, 'ymin': 0, 'xmax': 642, 'ymax': 476}}, ], ], ) @require_tf @unittest.skip('Zero Shot Object Detection not implemented in TF' ) def UpperCamelCase_ ( self ): '''simple docstring''' pass @require_torch @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : str = 0.2 SCREAMING_SNAKE_CASE : Union[str, Any] = pipeline('zero-shot-object-detection' ) SCREAMING_SNAKE_CASE : Any = object_detector( 'http://images.cocodataset.org/val2017/000000039769.jpg', candidate_labels=['cat', 'remote', 'couch'], threshold=A, ) self.assertEqual( nested_simplify(A, decimals=4 ), [ {'score': 0.28_68, 'label': 'cat', 'box': {'xmin': 324, 'ymin': 20, 'xmax': 640, 'ymax': 373}}, {'score': 0.2_77, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 72, 'xmax': 177, 'ymax': 115}}, {'score': 0.25_37, 'label': 'cat', 'box': {'xmin': 1, 'ymin': 55, 'xmax': 315, 'ymax': 472}}, ], ) @require_torch @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = 2 SCREAMING_SNAKE_CASE : List[str] = pipeline('zero-shot-object-detection' ) SCREAMING_SNAKE_CASE : int = object_detector( 'http://images.cocodataset.org/val2017/000000039769.jpg', candidate_labels=['cat', 'remote', 'couch'], top_k=A, ) self.assertEqual( nested_simplify(A, decimals=4 ), [ {'score': 0.28_68, 'label': 'cat', 'box': {'xmin': 324, 'ymin': 20, 'xmax': 640, 'ymax': 373}}, {'score': 0.2_77, 'label': 'remote', 'box': {'xmin': 40, 'ymin': 72, 'xmax': 177, 'ymax': 115}}, ], )
251
'''simple docstring''' from __future__ import annotations from scipy.special import comb # type: ignore class _a : '''simple docstring''' def __init__( self, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. SCREAMING_SNAKE_CASE : Optional[int] = len(A ) - 1 def UpperCamelCase_ ( self, A ): '''simple docstring''' assert 0 <= t <= 1, "Time t must be between 0 and 1." SCREAMING_SNAKE_CASE : list[float] = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree, A ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(A ), 5 ) == 1 return output_values def UpperCamelCase_ ( self, A ): '''simple docstring''' assert 0 <= t <= 1, "Time t must be between 0 and 1." SCREAMING_SNAKE_CASE : str = self.basis_function(A ) SCREAMING_SNAKE_CASE : str = 0.0 SCREAMING_SNAKE_CASE : List[Any] = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def UpperCamelCase_ ( self, A = 0.01 ): '''simple docstring''' from matplotlib import pyplot as plt # type: ignore SCREAMING_SNAKE_CASE : list[float] = [] # x coordinates of points to plot SCREAMING_SNAKE_CASE : list[float] = [] # y coordinates of points to plot SCREAMING_SNAKE_CASE : List[str] = 0.0 while t <= 1: SCREAMING_SNAKE_CASE : Optional[int] = self.bezier_curve_function(A ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size SCREAMING_SNAKE_CASE : List[Any] = [i[0] for i in self.list_of_points] SCREAMING_SNAKE_CASE : Union[str, Any] = [i[1] for i in self.list_of_points] plt.plot( A, A, color='blue', label='Curve of Degree ' + str(self.degree ), ) plt.scatter(A, A, color='red', label='Control Points' ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
251
1
import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin snake_case_ = get_tests_dir('''fixtures/test_sentencepiece.model''') snake_case_ = get_tests_dir('''fixtures/test_sentencepiece_bpe.model''') snake_case_ = '''pt''' if is_torch_available() else '''tf''' @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ (__snake_case , unittest.TestCase ): __lowerCamelCase : Union[str, Any] = CamembertTokenizer __lowerCamelCase : Dict = CamembertTokenizerFast __lowerCamelCase : Optional[int] = True __lowerCamelCase : List[Any] = True def snake_case_ ( self): super().setUp() # We have a SentencePiece fixture for testing lowercase__ : Any = CamembertTokenizer(a) tokenizer.save_pretrained(self.tmpdirname) def snake_case_ ( self): lowercase__ : Dict = '<pad>' lowercase__ : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a) , a) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a) , a) def snake_case_ ( self): lowercase__ : str = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<s>NOTUSED') self.assertEqual(vocab_keys[1] , '<pad>') self.assertEqual(vocab_keys[-1] , '<mask>') self.assertEqual(len(a) , 1004) def snake_case_ ( self): self.assertEqual(self.get_tokenizer().vocab_size , 1005) def snake_case_ ( self): lowercase__ : Any = CamembertTokenizer(a) tokenizer.save_pretrained(self.tmpdirname) lowercase__ : Tuple = CamembertTokenizerFast.from_pretrained(self.tmpdirname) lowercase__ : Optional[Any] = 'I was born in 92000, and this is falsé.' lowercase__ : int = tokenizer.encode(a) lowercase__ : Optional[Any] = rust_tokenizer.encode(a) self.assertListEqual(a , a) lowercase__ : str = tokenizer.encode(a , add_special_tokens=a) lowercase__ : int = rust_tokenizer.encode(a , add_special_tokens=a) self.assertListEqual(a , a) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) lowercase__ : Optional[int] = tokenizer.convert_ids_to_tokens(a) lowercase__ : Tuple = rust_tokenizer.tokenize(a) self.assertListEqual(a , a) def snake_case_ ( self): if not self.test_rust_tokenizer: return lowercase__ : str = self.get_tokenizer() lowercase__ : str = self.get_rust_tokenizer() lowercase__ : Any = 'I was born in 92000, and this is falsé.' lowercase__ : Union[str, Any] = tokenizer.tokenize(a) lowercase__ : Any = rust_tokenizer.tokenize(a) self.assertListEqual(a , a) lowercase__ : str = tokenizer.encode(a , add_special_tokens=a) lowercase__ : List[Any] = rust_tokenizer.encode(a , add_special_tokens=a) self.assertListEqual(a , a) lowercase__ : Dict = self.get_rust_tokenizer() lowercase__ : Any = tokenizer.encode(a) lowercase__ : str = rust_tokenizer.encode(a) self.assertListEqual(a , a) @slow def snake_case_ ( self): # fmt: off lowercase__ : Dict = {'input_ids': [[5, 54, 7196, 297, 30, 23, 776, 18, 11, 3215, 3705, 8252, 22, 3164, 1181, 2116, 29, 16, 813, 25, 791, 3314, 20, 3446, 38, 2_7575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9088, 20, 1517, 8, 2_2804, 1_8818, 10, 38, 629, 607, 607, 142, 19, 7196, 867, 56, 1_0326, 24, 2267, 20, 416, 5072, 1_5612, 233, 734, 7, 2399, 27, 16, 3015, 1649, 7, 24, 20, 4338, 2399, 27, 13, 3400, 14, 13, 6189, 8, 930, 9, 6]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. lowercase__ : Dict = [ 'Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ' 'utilisé principalement dans le domaine du traitement automatique des langues (TAL).', 'À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ' 'pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ' 'telles que la traduction et la synthèse de texte.', ] self.tokenizer_integration_test_util( expected_encoding=a , model_name='camembert-base' , revision='3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf' , sequences=a , )
216
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class SCREAMING_SNAKE_CASE__ (__snake_case ): def __init__( self , a , a = None , a = None , a = True , a = None , a = False , a = None , a = True , a = "arrow" , **a , ): super().__init__( split=a , features=a , cache_dir=a , keep_in_memory=a , streaming=a , **a , ) lowercase__ : Optional[int] = load_from_cache_file lowercase__ : Optional[int] = file_format lowercase__ : int = Spark( df=a , features=a , cache_dir=a , working_dir=a , **a , ) def snake_case_ ( self): if self.streaming: return self.builder.as_streaming_dataset(split=self.split) lowercase__ : Dict = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=a , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split)
216
1
import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py __a = 'src/diffusers' # Matches is_xxx_available() __a = re.compile(r'is\_([a-z_]*)_available\(\)') # Matches from xxx import bla __a = re.compile(r'\s+from\s+\S*\s+import\s+([^\(\s].*)\n') __a = '\n{0} = None\n' __a = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n\n @classmethod\n def from_config(cls, *args, **kwargs):\n requires_backends(cls, {1})\n\n @classmethod\n def from_pretrained(cls, *args, **kwargs):\n requires_backends(cls, {1})\n' __a = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' def a ( snake_case__: Optional[Any] ): '''simple docstring''' lowercase_ = _re_backend.findall(snake_case__ ) if len(snake_case__ ) == 0: return None return "_and_".join(snake_case__ ) def a ( ): '''simple docstring''' with open(os.path.join(snake_case__ , '''__init__.py''' ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ = f.readlines() # Get to the point we do the actual imports for type checking lowercase_ = 0 lowercase_ = {} # Go through the end of the file while line_index < len(snake_case__ ): # If the line contains is_backend_available, we grab all objects associated with the `else` block lowercase_ = find_backend(lines[line_index] ) if backend is not None: while not lines[line_index].startswith('''else:''' ): line_index += 1 line_index += 1 lowercase_ = [] # Until we unindent, add backend objects to the list while line_index < len(snake_case__ ) and len(lines[line_index] ) > 1: lowercase_ = lines[line_index] lowercase_ = _re_single_line_import.search(snake_case__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 8 ): objects.append(line[8:-2] ) line_index += 1 if len(snake_case__ ) > 0: lowercase_ = objects else: line_index += 1 return backend_specific_objects def a ( snake_case__: Optional[int] , snake_case__: List[Any] ): '''simple docstring''' if name.isupper(): return DUMMY_CONSTANT.format(snake_case__ ) elif name.islower(): return DUMMY_FUNCTION.format(snake_case__ , snake_case__ ) else: return DUMMY_CLASS.format(snake_case__ , snake_case__ ) def a ( snake_case__: List[str]=None ): '''simple docstring''' if backend_specific_objects is None: lowercase_ = read_init() # For special correspondence backend to module name as used in the function requires_modulename lowercase_ = {} for backend, objects in backend_specific_objects.items(): lowercase_ = '''[''' + ''', '''.join(F'''"{b}"''' for b in backend.split('''_and_''' ) ) + ''']''' lowercase_ = '''# This file is autogenerated by the command `make fix-copies`, do not edit.\n''' dummy_file += "from ..utils import DummyObject, requires_backends\n\n" dummy_file += "\n".join([create_dummy_object(snake_case__ , snake_case__ ) for o in objects] ) lowercase_ = dummy_file return dummy_files def a ( snake_case__: Optional[int]=False ): '''simple docstring''' lowercase_ = create_dummy_files() # For special correspondence backend to shortcut as used in utils/dummy_xxx_objects.py lowercase_ = {'''torch''': '''pt'''} # Locate actual dummy modules and read their content. lowercase_ = os.path.join(snake_case__ , '''utils''' ) lowercase_ = { backend: os.path.join(snake_case__ , F'''dummy_{short_names.get(snake_case__ , snake_case__ )}_objects.py''' ) for backend in dummy_files.keys() } lowercase_ = {} for backend, file_path in dummy_file_paths.items(): if os.path.isfile(snake_case__ ): with open(snake_case__ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowercase_ = f.read() else: lowercase_ = '''''' for backend in dummy_files.keys(): if dummy_files[backend] != actual_dummies[backend]: if overwrite: print( F'''Updating diffusers.utils.dummy_{short_names.get(snake_case__ , snake_case__ )}_objects.py as the main ''' '''__init__ has new objects.''' ) with open(dummy_file_paths[backend] , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(dummy_files[backend] ) else: raise ValueError( '''The main __init__ has objects that are not present in ''' F'''diffusers.utils.dummy_{short_names.get(snake_case__ , snake_case__ )}_objects.py. Run `make fix-copies` ''' '''to fix this.''' ) if __name__ == "__main__": __a = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') __a = parser.parse_args() check_dummies(args.fix_and_overwrite)
30
def a ( snake_case__: list ): '''simple docstring''' if len(snake_case__ ) <= 1: return [tuple(snake_case__ )] lowercase_ = [] def generate(snake_case__: int , snake_case__: list ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , snake_case__ ) for i in range(k - 1 ): if k % 2 == 0: # k is even lowercase_ , lowercase_ = arr[k - 1], arr[i] else: # k is odd lowercase_ , lowercase_ = arr[k - 1], arr[0] generate(k - 1 , snake_case__ ) generate(len(snake_case__ ) , snake_case__ ) return res if __name__ == "__main__": __a = input('Enter numbers separated by a comma:\n').strip() __a = [int(item) for item in user_input.split(',')] print(heaps(arr))
30
1
import argparse import re from typing import Dict import torch from datasets import Audio, Dataset, load_dataset, load_metric from transformers import AutoFeatureExtractor, pipeline def lowerCamelCase_ ( _a : Dataset , _a : Dict[str, str] ): '''simple docstring''' UpperCAmelCase_ : Optional[int] = args.log_outputs UpperCAmelCase_ : List[str] = """_""".join(args.dataset.split("""/""" ) + [args.config, args.split] ) # load metric UpperCAmelCase_ : int = load_metric("""wer""" ) UpperCAmelCase_ : Tuple = load_metric("""cer""" ) # compute metrics UpperCAmelCase_ : List[str] = wer.compute(references=result["""target"""] , predictions=result["""prediction"""] ) UpperCAmelCase_ : Optional[Any] = cer.compute(references=result["""target"""] , predictions=result["""prediction"""] ) # print & log results UpperCAmelCase_ : Any = F'''WER: {wer_result}\nCER: {cer_result}''' print(_a ) with open(F'''{dataset_id}_eval_results.txt''' , """w""" ) as f: f.write(_a ) # log all results in text file. Possibly interesting for analysis if log_outputs is not None: UpperCAmelCase_ : List[Any] = F'''log_{dataset_id}_predictions.txt''' UpperCAmelCase_ : Optional[Any] = F'''log_{dataset_id}_targets.txt''' with open(_a , """w""" ) as p, open(_a , """w""" ) as t: # mapping function to write output def write_to_file(_a : List[Any] , _a : Optional[int] ): p.write(F'''{i}''' + """\n""" ) p.write(batch["""prediction"""] + """\n""" ) t.write(F'''{i}''' + """\n""" ) t.write(batch["""target"""] + """\n""" ) result.map(_a , with_indices=_a ) def lowerCamelCase_ ( _a : str ): '''simple docstring''' UpperCAmelCase_ : int = """[,?.!\-\;\:\"“%‘”�—’…–]""" # noqa: W605 IMPORTANT: this should correspond to the chars that were ignored during training UpperCAmelCase_ : Optional[int] = re.sub(_a , """""" , text.lower() ) # In addition, we can normalize the target text, e.g. removing new lines characters etc... # note that order is important here! UpperCAmelCase_ : Optional[int] = ["""\n\n""", """\n""", """ """, """ """] for t in token_sequences_to_ignore: UpperCAmelCase_ : List[Any] = """ """.join(text.split(_a ) ) return text def lowerCamelCase_ ( _a : List[Any] ): '''simple docstring''' UpperCAmelCase_ : List[str] = load_dataset(args.dataset , args.config , split=args.split , use_auth_token=_a ) # for testing: only process the first two examples as a test # dataset = dataset.select(range(10)) # load processor UpperCAmelCase_ : List[Any] = AutoFeatureExtractor.from_pretrained(args.model_id ) UpperCAmelCase_ : str = feature_extractor.sampling_rate # resample audio UpperCAmelCase_ : int = dataset.cast_column("""audio""" , Audio(sampling_rate=_a ) ) # load eval pipeline if args.device is None: UpperCAmelCase_ : Any = 0 if torch.cuda.is_available() else -1 UpperCAmelCase_ : Union[str, Any] = pipeline("""automatic-speech-recognition""" , model=args.model_id , device=args.device ) # map function to decode audio def map_to_pred(_a : Dict ): UpperCAmelCase_ : Tuple = asr( batch["""audio"""]["""array"""] , chunk_length_s=args.chunk_length_s , stride_length_s=args.stride_length_s ) UpperCAmelCase_ : List[Any] = prediction["""text"""] UpperCAmelCase_ : int = normalize_text(batch["""sentence"""] ) return batch # run inference on all examples UpperCAmelCase_ : Optional[Any] = dataset.map(_a , remove_columns=dataset.column_names ) # compute and log_results # do not change function below log_results(_a , _a ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument( '''--model_id''', type=str, required=True, help='''Model identifier. Should be loadable with 🤗 Transformers''' ) parser.add_argument( '''--dataset''', type=str, required=True, help='''Dataset name to evaluate the `model_id`. Should be loadable with 🤗 Datasets''', ) parser.add_argument( '''--config''', type=str, required=True, help='''Config of the dataset. *E.g.* `\'en\'` for Common Voice''' ) parser.add_argument('''--split''', type=str, required=True, help='''Split of the dataset. *E.g.* `\'test\'`''') parser.add_argument( '''--chunk_length_s''', type=float, default=None, help='''Chunk length in seconds. Defaults to 5 seconds.''' ) parser.add_argument( '''--stride_length_s''', type=float, default=None, help='''Stride of the audio chunks. Defaults to 1 second.''' ) parser.add_argument( '''--log_outputs''', action='''store_true''', help='''If defined, write outputs to log file for analysis.''' ) parser.add_argument( '''--device''', type=int, default=None, help='''The device to run the pipeline on. -1 for CPU (default), 0 for the first GPU and so on.''', ) UpperCamelCase_ = parser.parse_args() main(args)
59
from scipy.stats import spearmanr import datasets UpperCamelCase_ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' UpperCamelCase_ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' UpperCamelCase_ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): '''simple docstring''' def A__ ( self: int ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { """predictions""": datasets.Value("""float""" ), """references""": datasets.Value("""float""" ), } ) ,reference_urls=["""https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"""] ,) def A__ ( self: int ,lowerCamelCase_: int ,lowerCamelCase_: List[Any] ,lowerCamelCase_: List[str]=False ) -> Dict: UpperCAmelCase_ : List[str] = spearmanr(lowerCamelCase_ ,lowerCamelCase_ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
59
1
'''simple docstring''' import importlib.metadata import operator import re import sys from typing import Optional from packaging import version UpperCamelCase__ = { '''<''': operator.lt, '''<=''': operator.le, '''==''': operator.eq, '''!=''': operator.ne, '''>=''': operator.ge, '''>''': operator.gt, } def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: if got_ver is None or want_ver is None: raise ValueError( F"""Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider""" F""" reinstalling {pkg}.""" ) if not ops[op](version.parse(lowerCAmelCase__ ) , version.parse(lowerCAmelCase__ ) ): raise ImportError( F"""{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}""" ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ = None ) -> None: UpperCAmelCase__ : Optional[int] = F"""\n{hint}""" if hint is not None else '''''' # non-versioned check if re.match(R'''^[\w_\-\d]+$''' , lowerCAmelCase__ ): UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : List[str] = requirement, None, None else: UpperCAmelCase__ : Tuple = re.findall(R'''^([^!=<>\s]+)([\s!=<>]{1,2}.+)''' , lowerCAmelCase__ ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but''' F""" got {requirement}""" ) UpperCAmelCase__ , UpperCAmelCase__ : List[str] = match[0] UpperCAmelCase__ : Dict = want_full.split(''',''' ) # there could be multiple requirements UpperCAmelCase__ : Optional[int] = {} for w in want_range: UpperCAmelCase__ : Dict = re.findall(R'''^([\s!=<>]{1,2})(.+)''' , lowerCAmelCase__ ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23,''' F""" but got {requirement}""" ) UpperCAmelCase__ , UpperCAmelCase__ : Optional[int] = match[0] UpperCAmelCase__ : Optional[Any] = want_ver if op not in ops: raise ValueError(F"""{requirement}: need one of {list(ops.keys() )}, but got {op}""" ) # special case if pkg == "python": UpperCAmelCase__ : Tuple = '''.'''.join([str(lowerCAmelCase__ ) for x in sys.version_info[:3]] ) for op, want_ver in wanted.items(): _compare_versions(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return # check if any version is installed try: UpperCAmelCase__ : Tuple = importlib.metadata.version(lowerCAmelCase__ ) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( F"""The '{requirement}' distribution was not found and is required by this application. {hint}""" ) # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) def a__ ( lowerCAmelCase__ ) -> Any: UpperCAmelCase__ : List[str] = '''Try: pip install transformers -U or pip install -e \'.[dev]\' if you\'re working with git main''' return require_version(lowerCAmelCase__ , lowerCAmelCase__ )
181
'''simple docstring''' import json import os from functools import lru_cache from typing import List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all BART models at https://huggingface.co/models?filter=bart UpperCamelCase__ = { '''vocab_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/vocab.json''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/vocab.json''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json''', }, '''merges_file''': { '''facebook/bart-base''': '''https://huggingface.co/facebook/bart-base/resolve/main/merges.txt''', '''facebook/bart-large''': '''https://huggingface.co/facebook/bart-large/resolve/main/merges.txt''', '''facebook/bart-large-mnli''': '''https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt''', '''facebook/bart-large-cnn''': '''https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt''', '''facebook/bart-large-xsum''': '''https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt''', '''yjernite/bart_eli5''': '''https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt''', }, } UpperCamelCase__ = { '''facebook/bart-base''': 1_0_2_4, '''facebook/bart-large''': 1_0_2_4, '''facebook/bart-large-mnli''': 1_0_2_4, '''facebook/bart-large-cnn''': 1_0_2_4, '''facebook/bart-large-xsum''': 1_0_2_4, '''yjernite/bart_eli5''': 1_0_2_4, } @lru_cache() def a__ ( ) -> List[Any]: UpperCAmelCase__ : int = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) UpperCAmelCase__ : Optional[int] = bs[:] UpperCAmelCase__ : List[str] = 0 for b in range(2**8 ): if b not in bs: bs.append(lowerCAmelCase__ ) cs.append(2**8 + n ) n += 1 UpperCAmelCase__ : Any = [chr(lowerCAmelCase__ ) for n in cs] return dict(zip(lowerCAmelCase__ , lowerCAmelCase__ ) ) def a__ ( lowerCAmelCase__ ) -> Union[str, Any]: UpperCAmelCase__ : str = set() UpperCAmelCase__ : str = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ : Optional[int] = char return pairs class lowerCamelCase_ ( __a ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = ['input_ids', 'attention_mask'] def __init__( self : Optional[int] , _A : Optional[int] , _A : List[Any] , _A : int="replace" , _A : List[Any]="<s>" , _A : List[Any]="</s>" , _A : List[Any]="</s>" , _A : Optional[int]="<s>" , _A : List[str]="<unk>" , _A : List[str]="<pad>" , _A : Union[str, Any]="<mask>" , _A : Any=False , **_A : Dict , ): '''simple docstring''' UpperCAmelCase__ : Dict = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else bos_token UpperCAmelCase__ : Any = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else eos_token UpperCAmelCase__ : Optional[int] = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else sep_token UpperCAmelCase__ : Tuple = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else cls_token UpperCAmelCase__ : int = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else unk_token UpperCAmelCase__ : Optional[Any] = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase__ : Optional[int] = AddedToken(_A , lstrip=_A , rstrip=_A ) if isinstance(_A , _A ) else mask_token super().__init__( errors=_A , bos_token=_A , eos_token=_A , unk_token=_A , sep_token=_A , cls_token=_A , pad_token=_A , mask_token=_A , add_prefix_space=_A , **_A , ) with open(_A , encoding='''utf-8''' ) as vocab_handle: UpperCAmelCase__ : Optional[Any] = json.load(_A ) UpperCAmelCase__ : Any = {v: k for k, v in self.encoder.items()} UpperCAmelCase__ : List[str] = errors # how to handle errors in decoding UpperCAmelCase__ : str = bytes_to_unicode() UpperCAmelCase__ : Dict = {v: k for k, v in self.byte_encoder.items()} with open(_A , encoding='''utf-8''' ) as merges_handle: UpperCAmelCase__ : str = merges_handle.read().split('''\n''' )[1:-1] UpperCAmelCase__ : Union[str, Any] = [tuple(merge.split() ) for merge in bpe_merges] UpperCAmelCase__ : Optional[Any] = dict(zip(_A , range(len(_A ) ) ) ) UpperCAmelCase__ : Optional[int] = {} UpperCAmelCase__ : int = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions UpperCAmelCase__ : List[Any] = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property def lowercase_ ( self : int ): '''simple docstring''' return len(self.encoder ) def lowercase_ ( self : Tuple ): '''simple docstring''' return dict(self.encoder , **self.added_tokens_encoder ) def lowercase_ ( self : List[Any] , _A : Tuple ): '''simple docstring''' if token in self.cache: return self.cache[token] UpperCAmelCase__ : Optional[Any] = tuple(_A ) UpperCAmelCase__ : Dict = get_pairs(_A ) if not pairs: return token while True: UpperCAmelCase__ : Optional[Any] = min(_A , key=lambda _A : self.bpe_ranks.get(_A , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ : str = bigram UpperCAmelCase__ : int = [] UpperCAmelCase__ : Tuple = 0 while i < len(_A ): try: UpperCAmelCase__ : Optional[int] = word.index(_A , _A ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ : Tuple = j if word[i] == first and i < len(_A ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ : Optional[Any] = tuple(_A ) UpperCAmelCase__ : List[Any] = new_word if len(_A ) == 1: break else: UpperCAmelCase__ : Union[str, Any] = get_pairs(_A ) UpperCAmelCase__ : Optional[Any] = ''' '''.join(_A ) UpperCAmelCase__ : List[Any] = word return word def lowercase_ ( self : str , _A : str ): '''simple docstring''' UpperCAmelCase__ : List[Any] = [] for token in re.findall(self.pat , _A ): UpperCAmelCase__ : str = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_A ).split(''' ''' ) ) return bpe_tokens def lowercase_ ( self : List[str] , _A : Any ): '''simple docstring''' return self.encoder.get(_A , self.encoder.get(self.unk_token ) ) def lowercase_ ( self : int , _A : List[str] ): '''simple docstring''' return self.decoder.get(_A ) def lowercase_ ( self : Tuple , _A : Any ): '''simple docstring''' UpperCAmelCase__ : Any = ''''''.join(_A ) UpperCAmelCase__ : List[str] = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def lowercase_ ( self : int , _A : str , _A : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(_A ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ : Tuple = os.path.join( _A , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) UpperCAmelCase__ : Any = os.path.join( _A , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(_A , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_A , ensure_ascii=_A ) + '''\n''' ) UpperCAmelCase__ : Union[str, Any] = 0 with open(_A , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _A : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) UpperCAmelCase__ : List[str] = token_index writer.write(''' '''.join(_A ) + '''\n''' ) index += 1 return vocab_file, merge_file def lowercase_ ( self : str , _A : List[int] , _A : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ : List[str] = [self.cls_token_id] UpperCAmelCase__ : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowercase_ ( self : Optional[int] , _A : List[int] , _A : Optional[List[int]] = None , _A : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_A , token_ids_a=_A , already_has_special_tokens=_A ) if token_ids_a is None: return [1] + ([0] * len(_A )) + [1] return [1] + ([0] * len(_A )) + [1, 1] + ([0] * len(_A )) + [1] def lowercase_ ( self : Dict , _A : List[int] , _A : Optional[List[int]] = None ): '''simple docstring''' UpperCAmelCase__ : List[str] = [self.sep_token_id] UpperCAmelCase__ : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ ( self : Optional[Any] , _A : Any , _A : Dict=False , **_A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Any = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_A ) > 0 and not text[0].isspace()): UpperCAmelCase__ : Tuple = ''' ''' + text return (text, kwargs)
181
1
'''simple docstring''' def lowerCAmelCase_ ( snake_case__ = 1000 ): '''simple docstring''' A : Optional[int] = 2**power A : str = 0 while n: A : Tuple = r + n % 10, n // 10 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
368
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A ( __snake_case ): def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Any: """simple docstring""" super().__init__() # make sure scheduler can always be converted to DDIM A : Dict = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=SCREAMING_SNAKE_CASE , scheduler=SCREAMING_SNAKE_CASE ) @torch.no_grad() def __call__( self , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = 0.0 , SCREAMING_SNAKE_CASE = 50 , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = "pil" , SCREAMING_SNAKE_CASE = True , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" if isinstance(self.unet.config.sample_size , SCREAMING_SNAKE_CASE ): A : List[Any] = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: A : Optional[int] = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) and len(SCREAMING_SNAKE_CASE ) != batch_size: raise ValueError( F'You have passed a list of generators of length {len(SCREAMING_SNAKE_CASE )}, but requested an effective batch' F' size of {batch_size}. Make sure the batch size matches the length of the generators.' ) A : str = randn_tensor(SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(SCREAMING_SNAKE_CASE ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output A : Any = self.unet(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 A : int = self.scheduler.step( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , eta=SCREAMING_SNAKE_CASE , use_clipped_model_output=SCREAMING_SNAKE_CASE , generator=SCREAMING_SNAKE_CASE ).prev_sample A : Dict = (image / 2 + 0.5).clamp(0 , 1 ) A : Optional[int] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": A : int = self.numpy_to_pil(SCREAMING_SNAKE_CASE ) if not return_dict: return (image,) return ImagePipelineOutput(images=SCREAMING_SNAKE_CASE )
311
0
"""simple docstring""" from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _SCREAMING_SNAKE_CASE ( A__ ): UpperCAmelCase_ :Optional[int] = ["image_processor", "tokenizer"] UpperCAmelCase_ :Optional[int] = "Pix2StructImageProcessor" UpperCAmelCase_ :str = ("T5Tokenizer", "T5TokenizerFast") def __init__( self , __A , __A ) -> int: lowerCAmelCase_ :Union[str, Any] = False super().__init__(__A , __A ) def __call__( self , __A=None , __A = None , __A = True , __A = False , __A = None , __A = None , __A = 2048 , __A = 0 , __A = None , __A = None , __A = False , __A = False , __A = False , __A = False , __A = False , __A = True , __A = None , **__A , ) -> BatchEncoding: if images is None and text is None: raise ValueError("""You have to specify either images or text.""" ) # Get only text if images is None and not self.image_processor.is_vqa: lowerCAmelCase_ :Tuple = self.tokenizer lowerCAmelCase_ :Dict = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) return text_encoding if not self.image_processor.is_vqa: # add pixel_values lowerCAmelCase_ :List[str] = self.image_processor( __A , return_tensors=__A , max_patches=__A , **__A ) else: # add pixel_values and bbox lowerCAmelCase_ :int = self.image_processor( __A , return_tensors=__A , max_patches=__A , header_text=__A , **__A ) if text is not None and not self.image_processor.is_vqa: lowerCAmelCase_ :Union[str, Any] = self.tokenizer( text=__A , add_special_tokens=__A , padding=__A , truncation=__A , max_length=__A , stride=__A , pad_to_multiple_of=__A , return_attention_mask=__A , return_overflowing_tokens=__A , return_special_tokens_mask=__A , return_offsets_mapping=__A , return_token_type_ids=__A , return_length=__A , verbose=__A , return_tensors=__A , **__A , ) if "attention_mask" in text_encoding: lowerCAmelCase_ :List[str] = text_encoding.pop("""attention_mask""" ) if "input_ids" in text_encoding: lowerCAmelCase_ :List[str] = text_encoding.pop("""input_ids""" ) else: lowerCAmelCase_ :Optional[Any] = None if text_encoding is not None: encoding_image_processor.update(__A ) return encoding_image_processor def __lowerCAmelCase ( self , *__A , **__A ) -> Any: return self.tokenizer.batch_decode(*__A , **__A ) def __lowerCAmelCase ( self , *__A , **__A ) -> Dict: return self.tokenizer.decode(*__A , **__A ) @property def __lowerCAmelCase ( self ) -> Optional[int]: lowerCAmelCase_ :List[str] = self.tokenizer.model_input_names lowerCAmelCase_ :Tuple = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
84
"""simple docstring""" import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) class _SCREAMING_SNAKE_CASE ( A__ ): def __init__( self , __A ) -> Optional[Any]: super().__init__() lowerCAmelCase_ :int = nn.ModuleList(__A ) def __lowerCAmelCase ( self , __A , __A , __A , __A , __A , __A = None , __A = None , __A = None , __A = None , __A = False , __A = True , ) -> Union[ControlNetOutput, Tuple]: for i, (image, scale, controlnet) in enumerate(zip(__A , __A , self.nets ) ): lowerCAmelCase_ , lowerCAmelCase_ :List[Any] = controlnet( __A , __A , __A , __A , __A , __A , __A , __A , __A , __A , __A , ) # merge samples if i == 0: lowerCAmelCase_ , lowerCAmelCase_ :Tuple = down_samples, mid_sample else: lowerCAmelCase_ :str = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(__A , __A ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def __lowerCAmelCase ( self , __A , __A = True , __A = None , __A = False , __A = None , ) -> Optional[Any]: lowerCAmelCase_ :int = 0 lowerCAmelCase_ :Dict = save_directory for controlnet in self.nets: controlnet.save_pretrained( __A , is_main_process=__A , save_function=__A , safe_serialization=__A , variant=__A , ) idx += 1 lowerCAmelCase_ :Any = model_path_to_save + f"""_{idx}""" @classmethod def __lowerCAmelCase ( cls , __A , **__A ) -> List[Any]: lowerCAmelCase_ :int = 0 lowerCAmelCase_ :Dict = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... lowerCAmelCase_ :List[Any] = pretrained_model_path while os.path.isdir(__A ): lowerCAmelCase_ :Tuple = ControlNetModel.from_pretrained(__A , **__A ) controlnets.append(__A ) idx += 1 lowerCAmelCase_ :Dict = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(__A )} controlnets loaded from {pretrained_model_path}.""" ) if len(__A ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(__A )}. Expected at least {pretrained_model_path + "_0"}.""" ) return cls(__A )
84
1
"""simple docstring""" class _lowerCamelCase : def __init__( self : Any , UpperCamelCase : list[int] ) -> None: """simple docstring""" lowerCAmelCase__ : Dict = len(UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = [0] * len_array if len_array > 0: lowerCAmelCase__ : Optional[int] = array[0] for i in range(1 , UpperCamelCase ): lowerCAmelCase__ : Any = self.prefix_sum[i - 1] + array[i] def _lowerCAmelCase ( self : Optional[Any] , UpperCamelCase : int , UpperCamelCase : int ) -> int: """simple docstring""" if start == 0: return self.prefix_sum[end] return self.prefix_sum[end] - self.prefix_sum[start - 1] def _lowerCAmelCase ( self : List[str] , UpperCamelCase : int ) -> bool: """simple docstring""" lowerCAmelCase__ : List[str] = {0} for sum_item in self.prefix_sum: if sum_item - target_sum in sums: return True sums.add(UpperCamelCase ) return False if __name__ == "__main__": import doctest doctest.testmod()
212
"""simple docstring""" import tempfile import unittest from transformers import TaConfig, is_torch_available from transformers.testing_utils import ( require_sentencepiece, require_tokenizers, require_torch, slow, torch_device, ) from ...generation.test_utils import GenerationTesterMixin from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import AutoTokenizer, UMTaForConditionalGeneration, UMTaForQuestionAnswering, UMTaModel class _lowerCamelCase : def __init__( self : str , UpperCamelCase : int , UpperCamelCase : str=99 , UpperCamelCase : Optional[int]=13 , UpperCamelCase : Dict=7 , UpperCamelCase : List[Any]=9 , UpperCamelCase : Optional[int]=True , UpperCamelCase : Any=True , UpperCamelCase : Union[str, Any]=False , UpperCamelCase : Optional[int]=32 , UpperCamelCase : str=5 , UpperCamelCase : int=4 , UpperCamelCase : Optional[Any]=37 , UpperCamelCase : Tuple=8 , UpperCamelCase : Any=0.1 , UpperCamelCase : Union[str, Any]=0.002 , UpperCamelCase : List[Any]=1 , UpperCamelCase : Any=0 , UpperCamelCase : Optional[Any]=0 , UpperCamelCase : Dict=None , UpperCamelCase : str=None , ) -> Any: """simple docstring""" lowerCAmelCase__ : Optional[Any] = parent lowerCAmelCase__ : Union[str, Any] = batch_size lowerCAmelCase__ : List[str] = encoder_seq_length lowerCAmelCase__ : Any = decoder_seq_length # For common tests lowerCAmelCase__ : Union[str, Any] = self.decoder_seq_length lowerCAmelCase__ : List[Any] = is_training lowerCAmelCase__ : Optional[Any] = use_attention_mask lowerCAmelCase__ : str = use_labels lowerCAmelCase__ : Any = vocab_size lowerCAmelCase__ : Any = hidden_size lowerCAmelCase__ : Optional[int] = num_hidden_layers lowerCAmelCase__ : Any = num_attention_heads lowerCAmelCase__ : int = d_ff lowerCAmelCase__ : int = relative_attention_num_buckets lowerCAmelCase__ : Union[str, Any] = dropout_rate lowerCAmelCase__ : str = initializer_factor lowerCAmelCase__ : Tuple = eos_token_id lowerCAmelCase__ : List[str] = pad_token_id lowerCAmelCase__ : str = decoder_start_token_id lowerCAmelCase__ : Optional[Any] = None lowerCAmelCase__ : Dict = decoder_layers def _lowerCAmelCase ( self : Dict ) -> Union[str, Any]: """simple docstring""" return TaConfig.from_pretrained("""google/umt5-base""" ) def _lowerCAmelCase ( self : str , UpperCamelCase : Optional[int] , UpperCamelCase : Dict , UpperCamelCase : Optional[int] , UpperCamelCase : List[Any]=None , UpperCamelCase : List[Any]=None , UpperCamelCase : List[Any]=None , UpperCamelCase : int=None , UpperCamelCase : List[Any]=None , ) -> List[Any]: """simple docstring""" if attention_mask is None: lowerCAmelCase__ : Optional[Any] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: lowerCAmelCase__ : List[str] = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: lowerCAmelCase__ : str = torch.ones(config.num_hidden_layers , config.num_attention_heads , device=UpperCamelCase ) if decoder_head_mask is None: lowerCAmelCase__ : Optional[int] = torch.ones(config.num_decoder_layers , config.num_attention_heads , device=UpperCamelCase ) if cross_attn_head_mask is None: lowerCAmelCase__ : Optional[int] = torch.ones( config.num_decoder_layers , config.num_attention_heads , device=UpperCamelCase ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } def _lowerCAmelCase ( self : Any ) -> Dict: """simple docstring""" lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size, self.encoder_seq_length] , self.vocab_size ) lowerCAmelCase__ : Tuple = ids_tensor([self.batch_size, self.decoder_seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for NllbMoe the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input lowerCAmelCase__ : int = input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ : Optional[Any] = decoder_input_ids.clamp(self.pad_token_id + 1 ) lowerCAmelCase__ : Tuple = self.get_config() lowerCAmelCase__ : Dict = config.num_attention_heads lowerCAmelCase__ : Dict = self.prepare_inputs_dict(UpperCamelCase , UpperCamelCase , UpperCamelCase ) return config, input_dict def _lowerCAmelCase ( self : Dict ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() return config, inputs_dict def _lowerCAmelCase ( self : Dict ) -> List[str]: """simple docstring""" return TaConfig( vocab_size=1_66 , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def _lowerCAmelCase ( self : Optional[int] ) -> Optional[int]: """simple docstring""" return TaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , d_ff=self.d_ff , d_kv=self.hidden_size // self.num_attention_heads , num_layers=self.num_hidden_layers , num_decoder_layers=self.decoder_layers , num_heads=self.num_attention_heads , relative_attention_num_buckets=self.relative_attention_num_buckets , dropout_rate=self.dropout_rate , initializer_factor=self.initializer_factor , eos_token_id=self.eos_token_id , bos_token_id=self.pad_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.decoder_start_token_id , ) def _lowerCAmelCase ( self : List[Any] , UpperCamelCase : List[Any] , UpperCamelCase : Optional[int] , UpperCamelCase : List[str] , UpperCamelCase : Optional[int] , UpperCamelCase : Any , UpperCamelCase : Any , ) -> List[Any]: """simple docstring""" lowerCAmelCase__ : List[Any] = UMTaModel(config=UpperCamelCase ) model.to(UpperCamelCase ) model.eval() lowerCAmelCase__ : Optional[int] = model( input_ids=UpperCamelCase , decoder_input_ids=UpperCamelCase , attention_mask=UpperCamelCase , decoder_attention_mask=UpperCamelCase , ) lowerCAmelCase__ : Optional[int] = model(input_ids=UpperCamelCase , decoder_input_ids=UpperCamelCase ) lowerCAmelCase__ : List[Any] = result.last_hidden_state lowerCAmelCase__ : Any = result.past_key_values lowerCAmelCase__ : str = result.encoder_last_hidden_state self.parent.assertEqual(encoder_output.size() , (self.batch_size, self.encoder_seq_length, self.hidden_size) ) self.parent.assertEqual(decoder_output.size() , (self.batch_size, self.decoder_seq_length, self.hidden_size) ) # There should be `num_layers` key value embeddings stored in decoder_past self.parent.assertEqual(len(UpperCamelCase ) , config.num_layers ) # There should be a self attn key, a self attn value, a cross attn key and a cross attn value stored in each decoder_past tuple self.parent.assertEqual(len(decoder_past[0] ) , 4 ) def _lowerCAmelCase ( self : Optional[int] , UpperCamelCase : List[str] , UpperCamelCase : int , UpperCamelCase : Optional[int] , UpperCamelCase : Any , UpperCamelCase : Optional[int] , UpperCamelCase : Optional[int] , ) -> Tuple: """simple docstring""" lowerCAmelCase__ : Dict = UMTaModel(config=UpperCamelCase ).get_decoder().to(UpperCamelCase ).eval() # first forward pass lowerCAmelCase__ : Optional[int] = model(UpperCamelCase , use_cache=UpperCamelCase ) lowerCAmelCase__ : Any = model(UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = model(UpperCamelCase , use_cache=UpperCamelCase ) self.parent.assertTrue(len(UpperCamelCase ) == len(UpperCamelCase ) ) self.parent.assertTrue(len(UpperCamelCase ) == len(UpperCamelCase ) + 1 ) lowerCAmelCase__ , lowerCAmelCase__ : List[str] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCAmelCase__ : Any = ids_tensor((self.batch_size, 1) , config.vocab_size ) # append to next input_ids and lowerCAmelCase__ : List[str] = torch.cat([input_ids, next_tokens] , dim=-1 ) lowerCAmelCase__ : List[str] = model(UpperCamelCase )["""last_hidden_state"""] lowerCAmelCase__ : Any = model(UpperCamelCase , past_key_values=UpperCamelCase )["""last_hidden_state"""] # select random slice lowerCAmelCase__ : Optional[int] = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowerCAmelCase__ : List[Any] = output_from_no_past[:, -1, random_slice_idx].detach() lowerCAmelCase__ : Optional[int] = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(UpperCamelCase , UpperCamelCase , atol=1E-3 ) ) def _lowerCAmelCase ( self : Optional[Any] , UpperCamelCase : str , UpperCamelCase : int , ) -> Dict: """simple docstring""" lowerCAmelCase__ : Union[str, Any] = UMTaModel(config=UpperCamelCase ).to(UpperCamelCase ).half().eval() lowerCAmelCase__ : Dict = model(**UpperCamelCase )["""last_hidden_state"""] self.parent.assertFalse(torch.isnan(UpperCamelCase ).any().item() ) @require_torch class _lowerCamelCase ( a_ , a_ , a_ , unittest.TestCase ): _lowerCamelCase :List[Any] = ( (UMTaModel, UMTaForConditionalGeneration, UMTaForQuestionAnswering) if is_torch_available() else () ) _lowerCamelCase :List[str] = (UMTaForConditionalGeneration,) if is_torch_available() else () _lowerCamelCase :Optional[Any] = ( { "conversational": UMTaForConditionalGeneration, "feature-extraction": UMTaModel, "summarization": UMTaForConditionalGeneration, "text2text-generation": UMTaForConditionalGeneration, "translation": UMTaForConditionalGeneration, "question-answering": UMTaForQuestionAnswering, } if is_torch_available() else {} ) _lowerCamelCase :Dict = True _lowerCamelCase :Optional[Any] = False _lowerCamelCase :List[str] = False _lowerCamelCase :Dict = True _lowerCamelCase :str = True # The small UMT5 model needs higher percentages for CPU/MP tests _lowerCamelCase :Optional[int] = [0.8, 0.9] def _lowerCAmelCase ( self : Dict ) -> List[str]: """simple docstring""" lowerCAmelCase__ : Any = UMTaModelTester(self ) @unittest.skip("""Test has a segmentation fault on torch 1.8.0""" ) def _lowerCAmelCase ( self : Union[str, Any] ) -> str: """simple docstring""" lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() lowerCAmelCase__ : List[str] = UMTaModel(config_and_inputs[0] ).to(UpperCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: torch.onnx.export( UpperCamelCase , (config_and_inputs[1], config_and_inputs[3], config_and_inputs[2]) , f"""{tmpdirname}/t5_test.onnx""" , export_params=UpperCamelCase , opset_version=9 , input_names=["""input_ids""", """decoder_input_ids"""] , ) @unittest.skipIf(torch_device == """cpu""" , """Cant do half precision""" ) def _lowerCAmelCase ( self : Union[str, Any] ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model_fpaa_forward(*UpperCamelCase ) def _lowerCAmelCase ( self : int ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase__ : Optional[int] = ["""encoder_attentions""", """decoder_attentions""", """cross_attentions"""] lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() lowerCAmelCase__ : List[Any] = config_and_inputs[0] lowerCAmelCase__ : int = UMTaForConditionalGeneration(UpperCamelCase ).eval() model.to(UpperCamelCase ) lowerCAmelCase__ : List[Any] = { """head_mask""": torch.zeros(config.num_layers , config.num_heads , device=UpperCamelCase ), """decoder_head_mask""": torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCamelCase ), """cross_attn_head_mask""": torch.zeros(config.num_decoder_layers , config.num_heads , device=UpperCamelCase ), } for attn_name, (name, mask) in zip(UpperCamelCase , head_masking.items() ): lowerCAmelCase__ : Union[str, Any] = {name: mask} # Explicitly pass decoder_head_mask as it is required from T5 model when head_mask specified if name == "head_mask": lowerCAmelCase__ : Tuple = torch.ones( config.num_decoder_layers , config.num_heads , device=UpperCamelCase ) lowerCAmelCase__ : Union[str, Any] = model.generate( config_and_inputs[1]["""input_ids"""] , num_beams=1 , max_length=3 , output_attentions=UpperCamelCase , return_dict_in_generate=UpperCamelCase , **UpperCamelCase , ) # We check the state of decoder_attentions and cross_attentions just from the last step lowerCAmelCase__ : str = out[attn_name] if attn_name == attention_names[0] else out[attn_name][-1] self.assertEqual(sum([w.sum().item() for w in attn_weights] ) , 0.0 ) @unittest.skip("""Does not work on the tiny model as we keep hitting edge cases.""" ) def _lowerCAmelCase ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" pass @require_torch @require_sentencepiece @require_tokenizers class _lowerCamelCase ( unittest.TestCase ): @slow @unittest.skip( """Unless we stop stripping left and right by default for all special tokens, the expected ids obtained here will not match the original ones. Wait for https://github.com/huggingface/transformers/pull/23909 to be merged""" ) def _lowerCAmelCase ( self : List[Any] ) -> Optional[int]: """simple docstring""" lowerCAmelCase__ : Dict = UMTaForConditionalGeneration.from_pretrained("""google/umt5-small""" , return_dict=UpperCamelCase ).to(UpperCamelCase ) lowerCAmelCase__ : Dict = AutoTokenizer.from_pretrained("""google/umt5-small""" , use_fast=UpperCamelCase , legacy=UpperCamelCase ) lowerCAmelCase__ : int = [ """Bonjour monsieur <extra_id_0> bien <extra_id_1>.""", """No se como puedo <extra_id_0>.""", """This is the reason why we <extra_id_0> them.""", """The <extra_id_0> walks in <extra_id_1>, seats""", """A <extra_id_0> walks into a bar and orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>.""", ] lowerCAmelCase__ : Union[str, Any] = tokenizer(UpperCamelCase , return_tensors="""pt""" , padding=UpperCamelCase ).input_ids # fmt: off lowerCAmelCase__ : List[Any] = torch.tensor( [ [ 3_85_30, 21_07_03, 25_62_99, 14_10, 25_62_98, 2_74, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 8_26, 3_21, 6_71, 2_59_22, 25_62_99, 2_74, 1, 0,0, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 14_60, 3_39, 3_12, 1_90_14, 1_06_20, 7_58, 25_62_99, 23_55,2_74, 1, 0, 0, 0, 0, 0, 0,0, 0], [ 5_17, 25_62_99, 1_48_69, 2_81, 3_01, 25_62_98, 2_75, 11_99_83,1, 0, 0, 0, 0, 0, 0, 0,0, 0], [ 3_20, 25_62_99, 1_48_69, 2_81, 22_34, 2_89, 22_75, 3_33,6_13_91, 2_89, 25_62_98, 5_43, 25_62_97, 16_87_14, 3_29, 25_62_96,2_74, 1], ] ) # fmt: on torch.testing.assert_allclose(UpperCamelCase , UpperCamelCase ) lowerCAmelCase__ : Optional[Any] = model.generate(input_ids.to(UpperCamelCase ) ) lowerCAmelCase__ : int = [ """<pad><extra_id_0> et<extra_id_1> [eod] <extra_id_2><extra_id_55>.. [eod] 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 💐 <extra_id_56>ajšietosto<extra_id_56>lleux<extra_id_19><extra_id_6>ajšie</s>""", """<pad><extra_id_0>.<extra_id_1>.,<0x0A>...spech <0x0A><extra_id_20> <extra_id_21></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", """<pad><extra_id_0> are not going to be a part of the world. We are not going to be a part of<extra_id_1> and<extra_id_2><0x0A><extra_id_48>.<extra_id_48></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", """<pad><extra_id_0> door<extra_id_1>, the door<extra_id_2> 피해[/</s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", """<pad><extra_id_0>nyone who<extra_id_1> drink<extra_id_2> a<extra_id_3> alcohol<extra_id_4> A<extra_id_5> A. This<extra_id_6> I<extra_id_7><extra_id_52><extra_id_53></s><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad><pad>""", ] lowerCAmelCase__ : Any = tokenizer.batch_decode(UpperCamelCase ) self.assertEqual(UpperCamelCase , UpperCamelCase )
212
1
"""simple docstring""" import json import os import torch from diffusers import UNetaDModel os.makedirs("""hub/hopper-medium-v2/unet/hor32""", exist_ok=True) os.makedirs("""hub/hopper-medium-v2/unet/hor128""", exist_ok=True) os.makedirs("""hub/hopper-medium-v2/value_function""", exist_ok=True) def UpperCAmelCase__ ( lowerCAmelCase__ :Dict ) -> Dict: '''simple docstring''' if hor == 1_2_8: lowercase = ("""DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""") lowercase = (3_2, 1_2_8, 2_5_6) lowercase = ("""UpResnetBlock1D""", """UpResnetBlock1D""") elif hor == 3_2: lowercase = ("""DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""") lowercase = (3_2, 6_4, 1_2_8, 2_5_6) lowercase = ("""UpResnetBlock1D""", """UpResnetBlock1D""", """UpResnetBlock1D""") lowercase = torch.load(f'/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch' ) lowercase = model.state_dict() lowercase = { """down_block_types""": down_block_types, """block_out_channels""": block_out_channels, """up_block_types""": up_block_types, """layers_per_block""": 1, """use_timestep_embedding""": True, """out_block_type""": """OutConv1DBlock""", """norm_num_groups""": 8, """downsample_each_block""": False, """in_channels""": 1_4, """out_channels""": 1_4, """extra_in_channels""": 0, """time_embedding_type""": """positional""", """flip_sin_to_cos""": False, """freq_shift""": 1, """sample_size""": 6_5_5_3_6, """mid_block_type""": """MidResTemporalBlock1D""", """act_fn""": """mish""", } lowercase = UNetaDModel(**lowerCAmelCase__ ) print(f'length of state dict: {len(state_dict.keys() )}' ) print(f'length of value function dict: {len(hf_value_function.state_dict().keys() )}' ) lowercase = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): lowercase = state_dict.pop(lowerCAmelCase__ ) hf_value_function.load_state_dict(lowerCAmelCase__ ) torch.save(hf_value_function.state_dict() , f'hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin' ) with open(f'hub/hopper-medium-v2/unet/hor{hor}/config.json' , """w""" ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) def UpperCAmelCase__ ( ) -> Union[str, Any]: '''simple docstring''' lowercase = { """in_channels""": 1_4, """down_block_types""": ("""DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D""", """DownResnetBlock1D"""), """up_block_types""": (), """out_block_type""": """ValueFunction""", """mid_block_type""": """ValueFunctionMidBlock1D""", """block_out_channels""": (3_2, 6_4, 1_2_8, 2_5_6), """layers_per_block""": 1, """downsample_each_block""": True, """sample_size""": 6_5_5_3_6, """out_channels""": 1_4, """extra_in_channels""": 0, """time_embedding_type""": """positional""", """use_timestep_embedding""": True, """flip_sin_to_cos""": False, """freq_shift""": 1, """norm_num_groups""": 8, """act_fn""": """mish""", } lowercase = torch.load("""/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch""" ) lowercase = model lowercase = UNetaDModel(**lowerCAmelCase__ ) print(f'length of state dict: {len(state_dict.keys() )}' ) print(f'length of value function dict: {len(hf_value_function.state_dict().keys() )}' ) lowercase = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): lowercase = state_dict.pop(lowerCAmelCase__ ) hf_value_function.load_state_dict(lowerCAmelCase__ ) torch.save(hf_value_function.state_dict() , """hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin""" ) with open("""hub/hopper-medium-v2/value_function/config.json""" , """w""" ) as f: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) if __name__ == "__main__": unet(3_2) # unet(128) value_function()
197
"""simple docstring""" from __future__ import annotations from collections.abc import Callable __lowerCAmelCase : str =list[list[float | int]] def UpperCAmelCase__ ( lowerCAmelCase__ :Matrix , lowerCAmelCase__ :Matrix ) -> Matrix: '''simple docstring''' lowercase = len(lowerCAmelCase__ ) lowercase = [[0 for _ in range(size + 1 )] for _ in range(lowerCAmelCase__ )] lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 for row in range(lowerCAmelCase__ ): for col in range(lowerCAmelCase__ ): lowercase = matrix[row][col] lowercase = vector[row][0] lowercase = 0 lowercase = 0 while row < size and col < size: # pivoting lowercase = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowerCAmelCase__ , lowerCAmelCase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: lowercase , lowercase = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowerCAmelCase__ ): lowercase = augmented[rowa][col] / augmented[row][col] lowercase = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowerCAmelCase__ ): for row in range(lowerCAmelCase__ ): lowercase = augmented[row][col] / augmented[col][col] for cola in range(lowerCAmelCase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 1_0 )] for row in range(lowerCAmelCase__ ) ] def UpperCAmelCase__ ( lowerCAmelCase__ :list[int] ) -> Callable[[int], int]: '''simple docstring''' lowercase = len(lowerCAmelCase__ ) lowercase = [[0 for _ in range(lowerCAmelCase__ )] for _ in range(lowerCAmelCase__ )] lowercase = [[0] for _ in range(lowerCAmelCase__ )] lowercase = 42 lowercase = 42 lowercase = 42 lowercase = 42 for x_val, y_val in enumerate(lowerCAmelCase__ ): for col in range(lowerCAmelCase__ ): lowercase = (x_val + 1) ** (size - col - 1) lowercase = y_val lowercase = solve(lowerCAmelCase__ , lowerCAmelCase__ ) def interpolated_func(lowerCAmelCase__ :int ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowerCAmelCase__ ) ) return interpolated_func def UpperCAmelCase__ ( lowerCAmelCase__ :int ) -> int: '''simple docstring''' return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**1_0 ) def UpperCAmelCase__ ( lowerCAmelCase__ :Callable[[int], int] = question_function , lowerCAmelCase__ :int = 1_0 ) -> int: '''simple docstring''' lowercase = [func(lowerCAmelCase__ ) for x_val in range(1 , order + 1 )] lowercase = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] lowercase = 0 lowercase = 42 lowercase = 42 for poly in polynomials: lowercase = 1 while func(lowerCAmelCase__ ) == poly(lowerCAmelCase__ ): x_val += 1 ret += poly(lowerCAmelCase__ ) return ret if __name__ == "__main__": print(F"""{solution() = }""")
197
1
import warnings from .generation import TFGenerationMixin class __lowerCamelCase ( lowerCamelCase_ ): """simple docstring""" warnings.warn( "Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will " "be removed in Transformers v5. Import as `from transformers import TFGenerationMixin` instead." , lowerCamelCase_ , )
371
import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Any , __lowerCamelCase: List[str] , __lowerCamelCase: List[Any] ): '''simple docstring''' return params[F'{prefix}/{prefix}/relpos_bias/rel_embedding'][:, i, :] def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: List[Any] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: int , __lowerCamelCase: Any="attention" ): '''simple docstring''' lowercase_ = lowercase_ = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/key/kernel'][:, i, :, :] ) lowercase_ = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) lowercase_ = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/out/kernel'][:, i, :, :] ) lowercase_ = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) lowercase_ = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/query/kernel'][:, i, :, :] ) lowercase_ = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) lowercase_ = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/value/kernel'][:, i, :, :] ) lowercase_ = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Optional[Any] , __lowerCamelCase: str , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Optional[Any]=False ): '''simple docstring''' if split_mlp_wi: lowercase_ = params[F'{prefix}/{prefix}/mlp/wi_0/kernel'][:, i, :] lowercase_ = params[F'{prefix}/{prefix}/mlp/wi_1/kernel'][:, i, :] lowercase_ = (wi_a, wi_a) else: lowercase_ = params[F'{prefix}/{prefix}/mlp/wi/kernel'][:, i, :] lowercase_ = params[F'{prefix}/{prefix}/mlp/wo/kernel'][:, i, :] return wi, wo def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Optional[int] , __lowerCamelCase: Dict , __lowerCamelCase: int , __lowerCamelCase: Optional[Any] ): '''simple docstring''' return params[F'{prefix}/{prefix}/{layer_name}/scale'][:, i] def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: dict , *, __lowerCamelCase: int , __lowerCamelCase: bool , __lowerCamelCase: bool = False ): '''simple docstring''' lowercase_ = traverse_util.flatten_dict(variables["target"] ) lowercase_ = {"/".join(__lowerCamelCase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi lowercase_ = "encoder/encoder/mlp/wi_0/kernel" in old print("Split MLP:" , __lowerCamelCase ) lowercase_ = collections.OrderedDict() # Shared embeddings. lowercase_ = old["token_embedder/embedding"] # Encoder. for i in range(__lowerCamelCase ): # Block i, layer 0 (Self Attention). lowercase_ = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , "pre_attention_layer_norm" ) lowercase_ , lowercase_ , lowercase_ , lowercase_ = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , "attention" ) lowercase_ = layer_norm lowercase_ = k.T lowercase_ = o.T lowercase_ = q.T lowercase_ = v.T # Block i, layer 1 (MLP). lowercase_ = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , "pre_mlp_layer_norm" ) lowercase_ , lowercase_ = tax_mlp_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , __lowerCamelCase ) lowercase_ = layer_norm if split_mlp_wi: lowercase_ = wi[0].T lowercase_ = wi[1].T else: lowercase_ = wi.T lowercase_ = wo.T if scalable_attention: # convert the rel_embedding of each layer lowercase_ = tax_relpos_bias_lookup( __lowerCamelCase , __lowerCamelCase , "encoder" ).T lowercase_ = old["encoder/encoder_norm/scale"] if not scalable_attention: lowercase_ = tax_relpos_bias_lookup( __lowerCamelCase , 0 , "encoder" ).T lowercase_ = tax_relpos_bias_lookup( __lowerCamelCase , 0 , "decoder" ).T if not is_encoder_only: # Decoder. for i in range(__lowerCamelCase ): # Block i, layer 0 (Self Attention). lowercase_ = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "pre_self_attention_layer_norm" ) lowercase_ , lowercase_ , lowercase_ , lowercase_ = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "self_attention" ) lowercase_ = layer_norm lowercase_ = k.T lowercase_ = o.T lowercase_ = q.T lowercase_ = v.T # Block i, layer 1 (Cross Attention). lowercase_ = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "pre_cross_attention_layer_norm" ) lowercase_ , lowercase_ , lowercase_ , lowercase_ = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "encoder_decoder_attention" ) lowercase_ = layer_norm lowercase_ = k.T lowercase_ = o.T lowercase_ = q.T lowercase_ = v.T # Block i, layer 2 (MLP). lowercase_ = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "pre_mlp_layer_norm" ) lowercase_ , lowercase_ = tax_mlp_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , __lowerCamelCase ) lowercase_ = layer_norm if split_mlp_wi: lowercase_ = wi[0].T lowercase_ = wi[1].T else: lowercase_ = wi.T lowercase_ = wo.T if scalable_attention: # convert the rel_embedding of each layer lowercase_ = tax_relpos_bias_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" ).T lowercase_ = old["decoder/decoder_norm/scale"] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: lowercase_ = old["decoder/logits_dense/kernel"].T return new def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Dict , __lowerCamelCase: bool ): '''simple docstring''' lowercase_ = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: lowercase_ = state_dict["shared.weight"] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: lowercase_ = state_dict["shared.weight"] if "lm_head.weight" not in state_dict: # For old 1.0 models. print("Using shared word embeddings as lm_head." ) lowercase_ = state_dict["shared.weight"] return state_dict def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Dict , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: List[Any] , __lowerCamelCase: Any ): '''simple docstring''' lowercase_ = checkpoints.load_tax_checkpoint(__lowerCamelCase ) lowercase_ = convert_tax_to_pytorch( __lowerCamelCase , num_layers=config.num_layers , is_encoder_only=__lowerCamelCase , scalable_attention=__lowerCamelCase ) lowercase_ = make_state_dict(__lowerCamelCase , __lowerCamelCase ) model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: Dict , __lowerCamelCase: Optional[Any] , __lowerCamelCase: List[str] , __lowerCamelCase: bool = False , __lowerCamelCase: bool = False , ): '''simple docstring''' lowercase_ = MTaConfig.from_json_file(__lowerCamelCase ) print(F'Building PyTorch model from configuration: {config}' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: lowercase_ = UMTaEncoderModel(__lowerCamelCase ) else: lowercase_ = UMTaForConditionalGeneration(__lowerCamelCase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(__lowerCamelCase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowerCamelCase ) print("Done" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser(description="""Converts a native T5X checkpoint into a PyTorch checkpoint.""") # Required parameters parser.add_argument( """--t5x_checkpoint_path""", default=None, type=str, required=True, help="""Path to the T5X checkpoint.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_encoder_only""", action="""store_true""", help="""Check if the model is encoder-decoder model""", default=False ) parser.add_argument( """--scalable_attention""", action="""store_true""", help="""Whether the model uses scaled attention (umt5 model)""", default=False, ) SCREAMING_SNAKE_CASE__ = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
297
0