code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def __UpperCamelCase ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[str] ): __a : str = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg''' __a : Dict = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ).convert('''RGB''' ) __a : Union[str, Any] = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73) , (0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11) ), ] ) __a : List[str] = transform(lowerCAmelCase__ ).unsqueeze(0 ).to(lowerCAmelCase__ ) return image def __UpperCamelCase ( lowerCAmelCase__ : Union[str, Any] ): if "visual_encoder" in key: __a : Optional[Any] = re.sub('''visual_encoder*''' , '''vision_model.encoder''' , lowerCAmelCase__ ) if "blocks" in key: __a : int = re.sub(R'''blocks''' , '''layers''' , lowerCAmelCase__ ) if "attn" in key: __a : int = re.sub(R'''attn''' , '''self_attn''' , lowerCAmelCase__ ) if "norm1" in key: __a : Dict = re.sub(R'''norm1''' , '''layer_norm1''' , lowerCAmelCase__ ) if "norm2" in key: __a : Any = re.sub(R'''norm2''' , '''layer_norm2''' , lowerCAmelCase__ ) if "encoder.norm" in key: __a : Dict = re.sub(R'''encoder.norm''' , '''post_layernorm''' , lowerCAmelCase__ ) if "encoder.patch_embed.proj" in key: __a : Union[str, Any] = re.sub(R'''encoder.patch_embed.proj''' , '''embeddings.patch_embedding''' , lowerCAmelCase__ ) if "encoder.pos_embed" in key: __a : Any = re.sub(R'''encoder.pos_embed''' , '''embeddings.position_embedding''' , lowerCAmelCase__ ) if "encoder.cls_token" in key: __a : str = re.sub(R'''encoder.cls_token''' , '''embeddings.class_embedding''' , lowerCAmelCase__ ) if "self_attn" in key: __a : List[str] = re.sub(R'''self_attn.proj''' , '''self_attn.projection''' , lowerCAmelCase__ ) return key @torch.no_grad() def __UpperCamelCase ( lowerCAmelCase__ : int , lowerCAmelCase__ : Any=None ): if config_path is not None: __a : Dict = BlipConfig.from_pretrained(lowerCAmelCase__ ) else: __a : List[str] = BlipConfig(projection_dim=5_1_2 , text_config={} , vision_config={} ) __a : Dict = BlipForConditionalGeneration(lowerCAmelCase__ ).eval() __a : Optional[Any] = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth''' __a : List[str] = blip_decoder(pretrained=lowerCAmelCase__ , image_size=3_8_4 , vit='''base''' ) __a : str = pt_model.eval() __a : Union[str, Any] = pt_model.state_dict() for key in modified_state_dict.copy(): __a : Any = modified_state_dict.pop(lowerCAmelCase__ ) __a : Dict = rename_key(lowerCAmelCase__ ) __a : List[str] = value hf_model.load_state_dict(lowerCAmelCase__ ) __a : List[str] = 3_8_4 __a : Any = load_demo_image(image_size=lowerCAmelCase__ , device='''cpu''' ) __a : int = BertTokenizer.from_pretrained('''bert-base-uncased''' ) __a : Optional[Any] = tokenizer(['''a picture of'''] ).input_ids __a : Union[str, Any] = hf_model.generate(lowerCAmelCase__ , lowerCAmelCase__ ) assert out[0].tolist() == [3_0_5_2_2, 1_0_3_7, 3_8_6_1, 1_9_9_7, 1_0_3_7, 2_4_5_0, 3_5_6_4, 2_0_0_6, 1_9_9_6, 3_5_0_9, 2_0_0_7, 2_0_1_4, 3_8_9_9, 1_0_2] __a : int = hf_model.generate(lowerCAmelCase__ ) assert out[0].tolist() == [3_0_5_2_2, 1_0_3_7, 2_4_5_0, 3_5_6_4, 2_0_0_6, 1_9_9_6, 3_5_0_9, 2_0_0_7, 2_0_1_4, 3_8_9_9, 1_0_2] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(lowerCAmelCase__ ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' __a : Optional[Any] = ( '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth''' ) __a : str = blip_vqa(pretrained=lowerCAmelCase__ , image_size=lowerCAmelCase__ , vit='''base''' ) vqa_model.eval() __a : str = vqa_model.state_dict() for key in modified_state_dict.copy(): __a : Tuple = modified_state_dict.pop(lowerCAmelCase__ ) __a : str = rename_key(lowerCAmelCase__ ) __a : str = value __a : List[str] = BlipForQuestionAnswering(lowerCAmelCase__ ) hf_vqa_model.load_state_dict(lowerCAmelCase__ ) __a : int = ['''How many dogs are in this image?'''] __a : List[Any] = tokenizer(lowerCAmelCase__ , return_tensors='''pt''' ).input_ids __a : str = hf_vqa_model.generate(lowerCAmelCase__ , lowerCAmelCase__ ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + '''_vqa''' ) __a : Union[str, Any] = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth''' __a : int = blip_itm(pretrained=lowerCAmelCase__ , image_size=lowerCAmelCase__ , vit='''base''' ) itm_model.eval() __a : int = itm_model.state_dict() for key in modified_state_dict.copy(): __a : List[Any] = modified_state_dict.pop(lowerCAmelCase__ ) __a : Optional[Any] = rename_key(lowerCAmelCase__ ) __a : Dict = value __a : Dict = BlipForImageTextRetrieval(lowerCAmelCase__ ) __a : List[str] = ['''A picture of a woman with a dog sitting in a beach'''] __a : Tuple = tokenizer( lowerCAmelCase__ , return_tensors='''pt''' , padding='''max_length''' , truncation=lowerCAmelCase__ , max_length=3_5 , ).input_ids hf_itm_model.load_state_dict(lowerCAmelCase__ ) hf_itm_model.eval() __a : Union[str, Any] = hf_itm_model(lowerCAmelCase__ , lowerCAmelCase__ , use_itm_head=lowerCAmelCase__ ) __a : List[Any] = hf_itm_model(lowerCAmelCase__ , lowerCAmelCase__ , use_itm_head=lowerCAmelCase__ ) assert out[0].item() == 0.21_10_68_74_94_27_79_54 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.4_56_98_84_53_86_50_51_27 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + '''_itm''' ) if __name__ == "__main__": lowercase__ =argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') lowercase__ =parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
216
def __UpperCamelCase ( lowerCAmelCase__ : int = 5_0_0_0_0_0_0_0 ): __a : int = set() __a : str = int((limit - 2_4) ** (1 / 2) ) __a : int = set(range(3 , prime_square_limit + 1 , 2 ) ) primes.add(2 ) for p in range(3 , prime_square_limit + 1 , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , prime_square_limit + 1 , lowerCAmelCase__ ) ) ) for primea in primes: __a : Union[str, Any] = primea * primea for primea in primes: __a : Union[str, Any] = primea * primea * primea if square + cube >= limit - 1_6: break for primea in primes: __a : int = primea * primea * primea * primea __a : Union[str, Any] = square + cube + tetr if total >= limit: break ret.add(lowerCAmelCase__ ) return len(lowerCAmelCase__ ) if __name__ == "__main__": print(F"""{solution() = }""")
216
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_owlvit import OwlViTImageProcessor _snake_case : Union[str, Any] = logging.get_logger(__name__) class A ( _a ): def __init__( self : Dict , *lowerCAmelCase_ : List[Any] , **lowerCAmelCase_ : Optional[Any] ) -> None: """simple docstring""" warnings.warn( '''The class OwlViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use OwlViTImageProcessor instead.''' , lowerCAmelCase_ , ) super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ )
179
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _snake_case : int = {'configuration_yolos': ['YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'YolosConfig', 'YolosOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = ['YolosFeatureExtractor'] _snake_case : Optional[int] = ['YolosImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = [ 'YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST', 'YolosForObjectDetection', 'YolosModel', 'YolosPreTrainedModel', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _snake_case : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
179
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : Union[str, Any] = logging.get_logger(__name__) a : List[str] = { '''junnyu/roformer_chinese_small''': '''https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json''', '''junnyu/roformer_chinese_base''': '''https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json''', '''junnyu/roformer_chinese_char_small''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json''' ), '''junnyu/roformer_chinese_char_base''': ( '''https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json''' ), '''junnyu/roformer_small_discriminator''': ( '''https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json''' ), '''junnyu/roformer_small_generator''': ( '''https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json''' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class __UpperCamelCase ( a__ ): lowerCamelCase : Optional[int] ="""roformer""" def __init__( self , lowerCAmelCase__=5_0000 , lowerCAmelCase__=None , lowerCAmelCase__=768 , lowerCAmelCase__=12 , lowerCAmelCase__=12 , lowerCAmelCase__=3072 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=1536 , lowerCAmelCase__=2 , lowerCAmelCase__=0.02 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=0 , lowerCAmelCase__=False , lowerCAmelCase__=True , **lowerCAmelCase__ , ) -> List[Any]: super().__init__(pad_token_id=lowerCAmelCase__ , **lowerCAmelCase__ ) a : Any = vocab_size a : Any = hidden_size if embedding_size is None else embedding_size a : Optional[Any] = hidden_size a : List[str] = num_hidden_layers a : List[str] = num_attention_heads a : Optional[int] = hidden_act a : Any = intermediate_size a : str = hidden_dropout_prob a : Tuple = attention_probs_dropout_prob a : str = max_position_embeddings a : List[Any] = type_vocab_size a : Optional[Any] = initializer_range a : Dict = layer_norm_eps a : str = rotary_value a : str = use_cache class __UpperCamelCase ( a__ ): @property def __a ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": a : Tuple = {0: "batch", 1: "choice", 2: "sequence"} else: a : Any = {0: "batch", 1: "sequence"} a : Union[str, Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
105
"""simple docstring""" def _SCREAMING_SNAKE_CASE ( _lowercase : float , _lowercase : float ) ->float: '''simple docstring''' return price * (1 + tax_rate) if __name__ == "__main__": print(F'''{price_plus_tax(100, 0.25) = }''') print(F'''{price_plus_tax(125.50, 0.05) = }''')
105
1
"""simple docstring""" import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class _a : def __init__( self : str, lowerCAmelCase__ : Dict, lowerCAmelCase__ : str=3, lowerCAmelCase__ : List[str]=7, lowerCAmelCase__ : int=True, lowerCAmelCase__ : Optional[Any]=True, lowerCAmelCase__ : int=False, lowerCAmelCase__ : Dict=True, lowerCAmelCase__ : int=9_9, lowerCAmelCase__ : Tuple=3_2, lowerCAmelCase__ : Any=5, lowerCAmelCase__ : str=4, lowerCAmelCase__ : int=3_7, lowerCAmelCase__ : Union[str, Any]="gelu", lowerCAmelCase__ : Dict=0.1, lowerCAmelCase__ : str=0.1, lowerCAmelCase__ : str=5_1_2, lowerCAmelCase__ : Optional[Any]=1_6, lowerCAmelCase__ : Union[str, Any]=2, lowerCAmelCase__ : Any=0.02, lowerCAmelCase__ : int=3, lowerCAmelCase__ : Tuple=4, lowerCAmelCase__ : List[str]=None, ) -> List[Any]: '''simple docstring''' _UpperCamelCase : Optional[int] = parent _UpperCamelCase : Any = batch_size _UpperCamelCase : str = seq_length _UpperCamelCase : Dict = is_training _UpperCamelCase : Dict = use_input_mask _UpperCamelCase : int = use_token_type_ids _UpperCamelCase : str = use_labels _UpperCamelCase : List[str] = vocab_size _UpperCamelCase : str = hidden_size _UpperCamelCase : Dict = num_hidden_layers _UpperCamelCase : int = num_attention_heads _UpperCamelCase : Dict = intermediate_size _UpperCamelCase : Dict = hidden_act _UpperCamelCase : Optional[Any] = hidden_dropout_prob _UpperCamelCase : List[str] = attention_probs_dropout_prob _UpperCamelCase : Tuple = max_position_embeddings _UpperCamelCase : Union[str, Any] = type_vocab_size _UpperCamelCase : str = type_sequence_label_size _UpperCamelCase : Dict = initializer_range _UpperCamelCase : List[Any] = num_labels _UpperCamelCase : Any = num_choices _UpperCamelCase : str = scope def snake_case ( self : Union[str, Any] ) -> Any: '''simple docstring''' _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) _UpperCamelCase : int = None if self.use_input_mask: _UpperCamelCase : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : int = None _UpperCamelCase : List[Any] = None _UpperCamelCase : Any = None _UpperCamelCase : Dict = None if self.use_labels: _UpperCamelCase : Optional[Any] = ids_tensor([self.batch_size], self.type_sequence_label_size ) _UpperCamelCase : List[str] = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) _UpperCamelCase : Optional[int] = ids_tensor([self.batch_size], self.num_choices ) _UpperCamelCase : str = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case ( self : str ) -> List[Any]: '''simple docstring''' return FalconConfig( vocab_size=self.vocab_size, hidden_size=self.hidden_size, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, is_decoder=lowerCAmelCase__, initializer_range=self.initializer_range, pad_token_id=1, new_decoder_architecture=lowerCAmelCase__, ) def snake_case ( self : Union[str, Any], lowerCAmelCase__ : Dict, lowerCAmelCase__ : List[Any], lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase : Dict = FalconModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : Union[str, Any] = model(lowerCAmelCase__, attention_mask=lowerCAmelCase__ ) _UpperCamelCase : Union[str, Any] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : str, lowerCAmelCase__ : Any, lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Union[str, Any], lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Tuple, lowerCAmelCase__ : str, lowerCAmelCase__ : Dict, lowerCAmelCase__ : Tuple, lowerCAmelCase__ : Optional[Any], ) -> Optional[int]: '''simple docstring''' _UpperCamelCase : Any = True _UpperCamelCase : Dict = FalconModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : Union[str, Any] = model( lowerCAmelCase__, attention_mask=lowerCAmelCase__, encoder_hidden_states=lowerCAmelCase__, encoder_attention_mask=lowerCAmelCase__, ) _UpperCamelCase : Union[str, Any] = model( lowerCAmelCase__, attention_mask=lowerCAmelCase__, encoder_hidden_states=lowerCAmelCase__, ) _UpperCamelCase : Tuple = model(lowerCAmelCase__, attention_mask=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : Optional[Any], lowerCAmelCase__ : int, lowerCAmelCase__ : Dict, lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Dict, lowerCAmelCase__ : str, lowerCAmelCase__ : List[Any], lowerCAmelCase__ : List[str], lowerCAmelCase__ : Dict, lowerCAmelCase__ : Optional[Any], ) -> Optional[int]: '''simple docstring''' _UpperCamelCase : List[Any] = FalconForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : int = model(lowerCAmelCase__, attention_mask=lowerCAmelCase__, labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self : Any, lowerCAmelCase__ : Any, lowerCAmelCase__ : List[str], lowerCAmelCase__ : str, lowerCAmelCase__ : int, lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : List[Any], lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Optional[int], ) -> int: '''simple docstring''' _UpperCamelCase : int = True _UpperCamelCase : Optional[int] = True _UpperCamelCase : Union[str, Any] = FalconForCausalLM(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() # first forward pass _UpperCamelCase : Union[str, Any] = model( lowerCAmelCase__, attention_mask=lowerCAmelCase__, encoder_hidden_states=lowerCAmelCase__, encoder_attention_mask=lowerCAmelCase__, use_cache=lowerCAmelCase__, ) _UpperCamelCase : int = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _UpperCamelCase : Dict = ids_tensor((self.batch_size, 3), config.vocab_size ) _UpperCamelCase : Dict = ids_tensor((self.batch_size, 3), vocab_size=2 ) # append to next input_ids and _UpperCamelCase : Dict = torch.cat([input_ids, next_tokens], dim=-1 ) _UpperCamelCase : List[str] = torch.cat([input_mask, next_mask], dim=-1 ) _UpperCamelCase : Dict = model( lowerCAmelCase__, attention_mask=lowerCAmelCase__, encoder_hidden_states=lowerCAmelCase__, encoder_attention_mask=lowerCAmelCase__, output_hidden_states=lowerCAmelCase__, )['''hidden_states'''][0] _UpperCamelCase : Tuple = model( lowerCAmelCase__, attention_mask=lowerCAmelCase__, encoder_hidden_states=lowerCAmelCase__, encoder_attention_mask=lowerCAmelCase__, past_key_values=lowerCAmelCase__, output_hidden_states=lowerCAmelCase__, )['''hidden_states'''][0] # select random slice _UpperCamelCase : Dict = ids_tensor((1,), output_from_past.shape[-1] ).item() _UpperCamelCase : Optional[int] = output_from_no_past[:, -3:, random_slice_idx].detach() _UpperCamelCase : List[str] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCAmelCase__, lowerCAmelCase__, atol=1e-3 ) ) def snake_case ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase : Optional[int] = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) : Optional[Any] = config_and_inputs _UpperCamelCase : List[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class _a ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): UpperCamelCase = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) UpperCamelCase = (FalconForCausalLM,) if is_torch_available() else () UpperCamelCase = ( { '''feature-extraction''': FalconModel, '''text-classification''': FalconForSequenceClassification, '''text-generation''': FalconForCausalLM, '''question-answering''': FalconForQuestionAnswering, '''token-classification''': FalconForTokenClassification, '''zero-shot''': FalconForSequenceClassification, } if is_torch_available() else {} ) UpperCamelCase = False UpperCamelCase = False def snake_case ( self : Dict ) -> str: '''simple docstring''' _UpperCamelCase : str = FalconModelTester(self ) _UpperCamelCase : Optional[int] = ConfigTester(self, config_class=lowerCAmelCase__, hidden_size=3_7 ) def snake_case ( self : List[str] ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() def snake_case ( self : Optional[Any] ) -> Dict: '''simple docstring''' _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def snake_case ( self : int ) -> Optional[int]: '''simple docstring''' _UpperCamelCase , *_UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: _UpperCamelCase : List[str] = alibi self.model_tester.create_and_check_model(lowerCAmelCase__, *lowerCAmelCase__ ) def snake_case ( self : Any ) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : List[Any] = input_dict['''input_ids'''] _UpperCamelCase : List[str] = input_ids.ne(1 ).to(lowerCAmelCase__ ) _UpperCamelCase : Tuple = ids_tensor([self.model_tester.batch_size], self.model_tester.type_sequence_label_size ) _UpperCamelCase : Any = FalconForSequenceClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : int = model(lowerCAmelCase__, attention_mask=lowerCAmelCase__, labels=lowerCAmelCase__ ) self.assertEqual(result.logits.shape, (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case ( self : Dict ) -> List[Any]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase : str = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase : Optional[int] = 3 _UpperCamelCase : Union[str, Any] = '''single_label_classification''' _UpperCamelCase : Optional[Any] = input_dict['''input_ids'''] _UpperCamelCase : List[Any] = input_ids.ne(1 ).to(lowerCAmelCase__ ) _UpperCamelCase : Tuple = ids_tensor([self.model_tester.batch_size], self.model_tester.type_sequence_label_size ) _UpperCamelCase : Dict = FalconForSequenceClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : Dict = model(lowerCAmelCase__, attention_mask=lowerCAmelCase__, labels=lowerCAmelCase__ ) self.assertEqual(result.logits.shape, (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case ( self : int ) -> List[str]: '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase : Optional[int] = input_dict['''input_ids'''] _UpperCamelCase : str = FalconForCausalLM(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : int = model(lowerCAmelCase__, use_cache=lowerCAmelCase__ ) _UpperCamelCase : List[str] = input_ids.shape[0] _UpperCamelCase : Optional[Any] = model._convert_to_rw_cache(result.past_key_values ) _UpperCamelCase : List[str] = model._convert_cache_to_standard_format(lowerCAmelCase__, lowerCAmelCase__ ) for layer in range(len(lowerCAmelCase__ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def snake_case ( self : List[str] ) -> Dict: '''simple docstring''' _UpperCamelCase , _UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCamelCase : Optional[Any] = 3 _UpperCamelCase : List[str] = '''multi_label_classification''' _UpperCamelCase : Any = input_dict['''input_ids'''] _UpperCamelCase : Optional[Any] = input_ids.ne(1 ).to(lowerCAmelCase__ ) _UpperCamelCase : Optional[int] = ids_tensor( [self.model_tester.batch_size, config.num_labels], self.model_tester.type_sequence_label_size ).to(torch.float ) _UpperCamelCase : Optional[Any] = FalconForSequenceClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : Any = model(lowerCAmelCase__, attention_mask=lowerCAmelCase__, labels=lowerCAmelCase__ ) self.assertEqual(result.logits.shape, (self.model_tester.batch_size, self.model_tester.num_labels) ) def snake_case ( self : Optional[int] ) -> str: '''simple docstring''' for model_class in self.all_generative_model_classes: _UpperCamelCase , _UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(lowerCAmelCase__, '''use_cache''' ): return _UpperCamelCase : Optional[Any] = model_class(lowerCAmelCase__ ).to(lowerCAmelCase__ ) if "use_cache" not in inputs: _UpperCamelCase : Dict = True _UpperCamelCase : str = model(**lowerCAmelCase__ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return _UpperCamelCase : int = ( getattr(lowerCAmelCase__, '''decoder_layers''', lowerCAmelCase__ ) or getattr(lowerCAmelCase__, '''num_decoder_layers''', lowerCAmelCase__ ) or config.num_hidden_layers ) _UpperCamelCase : List[str] = getattr(lowerCAmelCase__, '''num_kv_heads''', config.num_attention_heads ) _UpperCamelCase : int = getattr(lowerCAmelCase__, '''d_model''', config.hidden_size ) _UpperCamelCase : Optional[int] = embed_dim // num_attention_heads _UpperCamelCase : int = outputs['''past_key_values'''] self.assertEqual(len(lowerCAmelCase__ ), lowerCAmelCase__ ) _UpperCamelCase , _UpperCamelCase : Optional[Any] = inputs['''input_ids'''].shape for i in range(lowerCAmelCase__ ): if config.new_decoder_architecture: _UpperCamelCase : List[str] = config.num_attention_heads elif config.multi_query: _UpperCamelCase : Optional[Any] = 1 self.assertEqual(len(past_kv[0] ), 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape, (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape, (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class _a ( unittest.TestCase ): @slow def snake_case ( self : Any ) -> Any: '''simple docstring''' _UpperCamelCase : Optional[Any] = AutoTokenizer.from_pretrained('''Rocketknight1/falcon-rw-1b''' ) _UpperCamelCase : Union[str, Any] = FalconForCausalLM.from_pretrained('''Rocketknight1/falcon-rw-1b''' ) model.eval() model.to(lowerCAmelCase__ ) _UpperCamelCase : Dict = tokenizer('''My favorite food is''', return_tensors='''pt''' ).to(lowerCAmelCase__ ) _UpperCamelCase : Dict = ( '''My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday.''' ) _UpperCamelCase : Any = model.generate(**lowerCAmelCase__, do_sample=lowerCAmelCase__, max_new_tokens=1_9 ) _UpperCamelCase : List[str] = tokenizer.batch_decode(lowerCAmelCase__ )[0] self.assertEqual(lowerCAmelCase__, lowerCAmelCase__ ) @slow def snake_case ( self : List[str] ) -> Optional[Any]: '''simple docstring''' for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: _UpperCamelCase : int = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) _UpperCamelCase : List[str] = FalconForCausalLM.from_pretrained(lowerCAmelCase__ ) model.eval() model.to(lowerCAmelCase__ ) _UpperCamelCase : Tuple = tokenizer('''My favorite food is''', return_tensors='''pt''' ).to(lowerCAmelCase__ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**lowerCAmelCase__, do_sample=lowerCAmelCase__, max_new_tokens=4 ) model.generate(**lowerCAmelCase__, do_sample=lowerCAmelCase__, max_new_tokens=4 ) model.generate(**lowerCAmelCase__, num_beams=2, max_new_tokens=4 ) @slow def snake_case ( self : str ) -> List[Any]: '''simple docstring''' with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: _UpperCamelCase : Union[str, Any] = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) _UpperCamelCase : List[str] = FalconForCausalLM.from_pretrained(lowerCAmelCase__ ) model.eval() model.to(device=lowerCAmelCase__ ) _UpperCamelCase : Dict = tokenizer('''My favorite food is''', return_tensors='''pt''' ).to(lowerCAmelCase__ ) # Test results are the same with and without cache _UpperCamelCase : Optional[Any] = model.generate(**lowerCAmelCase__, do_sample=lowerCAmelCase__, max_new_tokens=2_0, use_cache=lowerCAmelCase__ ) _UpperCamelCase : str = model.generate(**lowerCAmelCase__, do_sample=lowerCAmelCase__, max_new_tokens=2_0, use_cache=lowerCAmelCase__ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
128
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class _a : def __init__( self : Dict, lowerCAmelCase__ : Optional[Any], lowerCAmelCase__ : Optional[int]=1_3, lowerCAmelCase__ : Optional[Any]=7, lowerCAmelCase__ : Optional[Any]=True, lowerCAmelCase__ : Any=True, lowerCAmelCase__ : str=True, lowerCAmelCase__ : Any=9_9, lowerCAmelCase__ : Dict=3_2, lowerCAmelCase__ : List[Any]=5, lowerCAmelCase__ : Tuple=4, lowerCAmelCase__ : List[Any]=3_7, lowerCAmelCase__ : Tuple="gelu", lowerCAmelCase__ : Any=0.1, lowerCAmelCase__ : Optional[Any]=0.1, lowerCAmelCase__ : Dict=5_1_2, lowerCAmelCase__ : List[str]=1_6, lowerCAmelCase__ : Tuple=2, lowerCAmelCase__ : int=0.02, lowerCAmelCase__ : int=3, lowerCAmelCase__ : Optional[Any]=4, lowerCAmelCase__ : Dict=None, ) -> int: '''simple docstring''' _UpperCamelCase : Tuple = parent _UpperCamelCase : Union[str, Any] = batch_size _UpperCamelCase : Union[str, Any] = seq_length _UpperCamelCase : Tuple = is_training _UpperCamelCase : Tuple = use_token_type_ids _UpperCamelCase : Optional[int] = use_labels _UpperCamelCase : Dict = vocab_size _UpperCamelCase : int = hidden_size _UpperCamelCase : Optional[Any] = num_hidden_layers _UpperCamelCase : str = num_attention_heads _UpperCamelCase : Union[str, Any] = intermediate_size _UpperCamelCase : List[str] = hidden_act _UpperCamelCase : Optional[Any] = hidden_dropout_prob _UpperCamelCase : int = attention_probs_dropout_prob _UpperCamelCase : Union[str, Any] = max_position_embeddings _UpperCamelCase : int = type_vocab_size _UpperCamelCase : List[str] = type_sequence_label_size _UpperCamelCase : List[str] = initializer_range _UpperCamelCase : int = num_labels _UpperCamelCase : List[str] = num_choices _UpperCamelCase : str = scope _UpperCamelCase : Optional[int] = self.vocab_size - 1 def snake_case ( self : int ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) _UpperCamelCase : List[str] = None if self.use_token_type_ids: _UpperCamelCase : Tuple = ids_tensor([self.batch_size, self.seq_length], self.type_vocab_size ) _UpperCamelCase : Optional[int] = None _UpperCamelCase : str = None _UpperCamelCase : List[str] = None if self.use_labels: _UpperCamelCase : int = ids_tensor([self.batch_size], self.type_sequence_label_size ) _UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) _UpperCamelCase : Dict = ids_tensor([self.batch_size], self.num_choices ) _UpperCamelCase : str = OpenAIGPTConfig( vocab_size=self.vocab_size, n_embd=self.hidden_size, n_layer=self.num_hidden_layers, n_head=self.num_attention_heads, n_positions=self.max_position_embeddings, pad_token_id=self.pad_token_id, ) _UpperCamelCase : List[Any] = ids_tensor([self.num_hidden_layers, self.num_attention_heads], 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def snake_case ( self : Union[str, Any], lowerCAmelCase__ : Optional[int], lowerCAmelCase__ : List[str], lowerCAmelCase__ : List[str], lowerCAmelCase__ : List[str], *lowerCAmelCase__ : List[Any] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : Dict = OpenAIGPTModel(config=lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : List[str] = model(lowerCAmelCase__, token_type_ids=lowerCAmelCase__, head_mask=lowerCAmelCase__ ) _UpperCamelCase : Any = model(lowerCAmelCase__, token_type_ids=lowerCAmelCase__ ) _UpperCamelCase : List[str] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape, (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case ( self : Any, lowerCAmelCase__ : Tuple, lowerCAmelCase__ : Union[str, Any], lowerCAmelCase__ : Any, lowerCAmelCase__ : Optional[Any], *lowerCAmelCase__ : Union[str, Any] ) -> Any: '''simple docstring''' _UpperCamelCase : Any = OpenAIGPTLMHeadModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : Tuple = model(lowerCAmelCase__, token_type_ids=lowerCAmelCase__, labels=lowerCAmelCase__ ) self.parent.assertEqual(result.loss.shape, () ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self : Optional[int], lowerCAmelCase__ : str, lowerCAmelCase__ : List[str], lowerCAmelCase__ : Any, lowerCAmelCase__ : List[Any], *lowerCAmelCase__ : Any ) -> int: '''simple docstring''' _UpperCamelCase : Tuple = OpenAIGPTDoubleHeadsModel(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : Optional[int] = model(lowerCAmelCase__, token_type_ids=lowerCAmelCase__, labels=lowerCAmelCase__ ) self.parent.assertEqual(result.loss.shape, () ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case ( self : List[str], lowerCAmelCase__ : Dict, lowerCAmelCase__ : Dict, lowerCAmelCase__ : List[str], lowerCAmelCase__ : Optional[Any], *lowerCAmelCase__ : List[str] ) -> int: '''simple docstring''' _UpperCamelCase : List[Any] = self.num_labels _UpperCamelCase : Optional[int] = OpenAIGPTForSequenceClassification(lowerCAmelCase__ ) model.to(lowerCAmelCase__ ) model.eval() _UpperCamelCase : str = ids_tensor([self.batch_size], self.type_sequence_label_size ) _UpperCamelCase : Union[str, Any] = model(lowerCAmelCase__, token_type_ids=lowerCAmelCase__, labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape, (self.batch_size, self.num_labels) ) def snake_case ( self : str ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : Any = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) : Tuple = config_and_inputs _UpperCamelCase : Tuple = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''head_mask''': head_mask, } return config, inputs_dict @require_torch class _a ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): UpperCamelCase = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) UpperCamelCase = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly UpperCamelCase = ( { '''feature-extraction''': OpenAIGPTModel, '''text-classification''': OpenAIGPTForSequenceClassification, '''text-generation''': OpenAIGPTLMHeadModel, '''zero-shot''': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def snake_case ( self : Union[str, Any], lowerCAmelCase__ : Any, lowerCAmelCase__ : List[str], lowerCAmelCase__ : str, lowerCAmelCase__ : List[str], lowerCAmelCase__ : List[str] ) -> List[str]: '''simple docstring''' if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def snake_case ( self : str, lowerCAmelCase__ : Optional[int], lowerCAmelCase__ : List[str], lowerCAmelCase__ : Optional[int]=False ) -> Tuple: '''simple docstring''' _UpperCamelCase : Optional[Any] = super()._prepare_for_class(lowerCAmelCase__, lowerCAmelCase__, return_labels=lowerCAmelCase__ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": _UpperCamelCase : Union[str, Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length), dtype=torch.long, device=lowerCAmelCase__, ) _UpperCamelCase : Tuple = inputs_dict['''labels'''] _UpperCamelCase : List[str] = inputs_dict['''labels'''] _UpperCamelCase : Optional[Any] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices), dtype=torch.long, device=lowerCAmelCase__, ) _UpperCamelCase : Dict = torch.zeros( self.model_tester.batch_size, dtype=torch.long, device=lowerCAmelCase__ ) return inputs_dict def snake_case ( self : List[str] ) -> List[str]: '''simple docstring''' _UpperCamelCase : Optional[Any] = OpenAIGPTModelTester(self ) _UpperCamelCase : int = ConfigTester(self, config_class=lowerCAmelCase__, n_embd=3_7 ) def snake_case ( self : Optional[int] ) -> str: '''simple docstring''' self.config_tester.run_common_tests() def snake_case ( self : Optional[int] ) -> Any: '''simple docstring''' _UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*lowerCAmelCase__ ) def snake_case ( self : Any ) -> Dict: '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowerCAmelCase__ ) def snake_case ( self : int ) -> Dict: '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*lowerCAmelCase__ ) def snake_case ( self : List[str] ) -> int: '''simple docstring''' _UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*lowerCAmelCase__ ) @slow def snake_case ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase : int = OpenAIGPTModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @require_torch class _a ( unittest.TestCase ): @slow def snake_case ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' _UpperCamelCase : int = OpenAIGPTLMHeadModel.from_pretrained('''openai-gpt''' ) model.to(lowerCAmelCase__ ) _UpperCamelCase : str = torch.tensor([[4_8_1, 4_7_3_5, 5_4_4]], dtype=torch.long, device=lowerCAmelCase__ ) # the president is _UpperCamelCase : Optional[int] = [ 4_8_1, 4_7_3_5, 5_4_4, 2_4_6, 9_6_3, 8_7_0, 7_6_2, 2_3_9, 2_4_4, 4_0_4_7_7, 2_4_4, 2_4_9, 7_1_9, 8_8_1, 4_8_7, 5_4_4, 2_4_0, 2_4_4, 6_0_3, 4_8_1, ] # the president is a very good man. " \n " i\'m sure he is, " said the _UpperCamelCase : Union[str, Any] = model.generate(lowerCAmelCase__, do_sample=lowerCAmelCase__ ) self.assertListEqual(output_ids[0].tolist(), lowerCAmelCase__ )
128
1
"""simple docstring""" def _snake_case ( lowercase__ , lowercase__ ): return int((input_a, input_a).count(0 ) == 0 ) def _snake_case ( ): assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
96
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) # TODO Update this UpperCAmelCase__ = { '''facebook/esm-1b''': '''https://huggingface.co/facebook/esm-1b/resolve/main/config.json''', # See all ESM models at https://huggingface.co/models?filter=esm } class lowerCamelCase__ ( lowerCAmelCase): SCREAMING_SNAKE_CASE__ = '''esm''' def __init__(self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=7_6_8 , UpperCAmelCase=1_2 , UpperCAmelCase=1_2 , UpperCAmelCase=3_0_7_2 , UpperCAmelCase=0.1 , UpperCAmelCase=0.1 , UpperCAmelCase=1_0_2_6 , UpperCAmelCase=0.02 , UpperCAmelCase=1e-12 , UpperCAmelCase="absolute" , UpperCAmelCase=True , UpperCAmelCase=None , UpperCAmelCase=False , UpperCAmelCase=False , UpperCAmelCase=None , UpperCAmelCase=None , **UpperCAmelCase , ) -> Tuple: super().__init__(pad_token_id=UpperCAmelCase , mask_token_id=UpperCAmelCase , **UpperCAmelCase ) _lowercase =vocab_size _lowercase =hidden_size _lowercase =num_hidden_layers _lowercase =num_attention_heads _lowercase =intermediate_size _lowercase =hidden_dropout_prob _lowercase =attention_probs_dropout_prob _lowercase =max_position_embeddings _lowercase =initializer_range _lowercase =layer_norm_eps _lowercase =position_embedding_type _lowercase =use_cache _lowercase =emb_layer_norm_before _lowercase =token_dropout _lowercase =is_folding_model if is_folding_model: if esmfold_config is None: logger.info('''No esmfold_config supplied for folding model, using default values.''' ) _lowercase =EsmFoldConfig() elif isinstance(UpperCAmelCase , UpperCAmelCase ): _lowercase =EsmFoldConfig(**UpperCAmelCase ) _lowercase =esmfold_config if vocab_list is None: logger.warning('''No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!''' ) _lowercase =get_default_vocab_list() else: _lowercase =vocab_list else: _lowercase =None _lowercase =None if self.esmfold_config is not None and getattr(self.esmfold_config , '''use_esm_attn_map''' , UpperCAmelCase ): raise ValueError('''The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!''' ) def __A (self ) -> List[str]: _lowercase =super().to_dict() if isinstance(self.esmfold_config , UpperCAmelCase ): _lowercase =self.esmfold_config.to_dict() return output @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = True SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = None def __A (self ) -> Union[str, Any]: if self.trunk is None: _lowercase =TrunkConfig() elif isinstance(self.trunk , UpperCAmelCase ): _lowercase =TrunkConfig(**self.trunk ) def __A (self ) -> Tuple: _lowercase =asdict(self ) _lowercase =self.trunk.to_dict() return output @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE__ = 48 SCREAMING_SNAKE_CASE__ = 1024 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = 32 SCREAMING_SNAKE_CASE__ = 32 SCREAMING_SNAKE_CASE__ = 32 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = 0 SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = 4 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = None def __A (self ) -> List[str]: if self.structure_module is None: _lowercase =StructureModuleConfig() elif isinstance(self.structure_module , UpperCAmelCase ): _lowercase =StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(f"`max_recycles` should be positive, got {self.max_recycles}." ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( '''`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got''' f" {self.sequence_state_dim} and {self.sequence_state_dim}." ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( '''`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got''' f" {self.pairwise_state_dim} and {self.pairwise_state_dim}." ) _lowercase =self.sequence_state_dim // self.sequence_head_width _lowercase =self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( '''`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got''' f" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}." ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( '''`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got''' f" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}." ) if self.pairwise_state_dim % 2 != 0: raise ValueError(f"`pairwise_state_dim` should be even, got {self.pairwise_state_dim}." ) if self.dropout >= 0.4: raise ValueError(f"`dropout` should not be greater than 0.4, got {self.dropout}." ) def __A (self ) -> Dict: _lowercase =asdict(self ) _lowercase =self.structure_module.to_dict() return output @dataclass class lowerCamelCase__ : SCREAMING_SNAKE_CASE__ = 384 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = 16 SCREAMING_SNAKE_CASE__ = 128 SCREAMING_SNAKE_CASE__ = 12 SCREAMING_SNAKE_CASE__ = 4 SCREAMING_SNAKE_CASE__ = 8 SCREAMING_SNAKE_CASE__ = 0.1 SCREAMING_SNAKE_CASE__ = 8 SCREAMING_SNAKE_CASE__ = 1 SCREAMING_SNAKE_CASE__ = 2 SCREAMING_SNAKE_CASE__ = 7 SCREAMING_SNAKE_CASE__ = 10 SCREAMING_SNAKE_CASE__ = 1E-8 SCREAMING_SNAKE_CASE__ = 1E5 def __A (self ) -> List[Any]: return asdict(self ) def UpperCAmelCase_ ( ) -> Tuple: """simple docstring""" return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
5
0
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class A( UpperCamelCase ): '''simple docstring''' def a__ ( self : List[str] ) -> int: """simple docstring""" lowerCamelCase_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(A_ , 'tf_padding' ) ) self.parent.assertTrue(hasattr(A_ , 'depth_multiplier' ) ) class A: '''simple docstring''' def __init__( self : Dict , A_ : Optional[Any] , A_ : Tuple=13 , A_ : Any=3 , A_ : Union[str, Any]=32 , A_ : List[str]=0.25 , A_ : Optional[Any]=8 , A_ : str=8 , A_ : Union[str, Any]=6 , A_ : Optional[Any]=32 , A_ : Dict=True , A_ : Optional[Any]=True , A_ : List[Any]=True , A_ : Optional[int]="relu6" , A_ : List[str]=1280 , A_ : Optional[Any]=0.1 , A_ : List[Any]=0.02 , A_ : Any=True , A_ : Union[str, Any]=True , A_ : int=10 , A_ : str=None , ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = num_channels lowerCamelCase_ = image_size lowerCamelCase_ = depth_multiplier lowerCamelCase_ = depth_divisible_by lowerCamelCase_ = min_depth lowerCamelCase_ = expand_ratio lowerCamelCase_ = tf_padding lowerCamelCase_ = output_stride lowerCamelCase_ = first_layer_is_expansion lowerCamelCase_ = finegrained_output lowerCamelCase_ = hidden_act lowerCamelCase_ = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) lowerCamelCase_ = classifier_dropout_prob lowerCamelCase_ = use_labels lowerCamelCase_ = is_training lowerCamelCase_ = num_labels lowerCamelCase_ = initializer_range lowerCamelCase_ = scope def a__ ( self : List[Any] ) -> Dict: """simple docstring""" lowerCamelCase_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase_ = None lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase_ = self.get_config() return config, pixel_values, labels, pixel_labels def a__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def a__ ( self : Union[str, Any] , A_ : Tuple , A_ : Optional[Any] , A_ : Dict , A_ : Optional[int] ) -> int: """simple docstring""" lowerCamelCase_ = MobileNetVaModel(config=A_ ) model.to(A_ ) model.eval() lowerCamelCase_ = model(A_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def a__ ( self : List[str] , A_ : List[str] , A_ : Any , A_ : str , A_ : Optional[int] ) -> List[str]: """simple docstring""" lowerCamelCase_ = self.num_labels lowerCamelCase_ = MobileNetVaForImageClassification(A_ ) model.to(A_ ) model.eval() lowerCamelCase_ = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a__ ( self : List[str] , A_ : Tuple , A_ : Optional[int] , A_ : str , A_ : Optional[int] ) -> Tuple: """simple docstring""" lowerCamelCase_ = self.num_labels lowerCamelCase_ = MobileNetVaForSemanticSegmentation(A_ ) model.to(A_ ) model.eval() lowerCamelCase_ = model(A_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) lowerCamelCase_ = model(A_ , labels=A_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def a__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = self.prepare_config_and_inputs() lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = config_and_inputs lowerCamelCase_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A( UpperCamelCase , UpperCamelCase , unittest.TestCase ): '''simple docstring''' UpperCamelCase = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) UpperCamelCase = ( { '''feature-extraction''': MobileNetVaModel, '''image-classification''': MobileNetVaForImageClassification, '''image-segmentation''': MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False UpperCamelCase = False def a__ ( self : Tuple ) -> Tuple: """simple docstring""" lowerCamelCase_ = MobileNetVaModelTester(self ) lowerCamelCase_ = MobileNetVaConfigTester(self , config_class=A_ , has_text_modality=A_ ) def a__ ( self : Optional[int] ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV2 does not use inputs_embeds' ) def a__ ( self : Tuple ) -> Optional[int]: """simple docstring""" pass @unittest.skip(reason='MobileNetV2 does not support input and output embeddings' ) def a__ ( self : List[Any] ) -> Optional[Any]: """simple docstring""" pass @unittest.skip(reason='MobileNetV2 does not output attentions' ) def a__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" pass def a__ ( self : List[str] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(A_ ) lowerCamelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ = [*signature.parameters.keys()] lowerCamelCase_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , A_ ) def a__ ( self : Any ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def a__ ( self : Optional[int] ) -> Dict: """simple docstring""" def check_hidden_states_output(A_ : List[Any] , A_ : int , A_ : Union[str, Any] ): lowerCamelCase_ = model_class(A_ ) model.to(A_ ) model.eval() with torch.no_grad(): lowerCamelCase_ = model(**self._prepare_for_class(A_ , A_ ) ) lowerCamelCase_ = outputs.hidden_states lowerCamelCase_ = 16 self.assertEqual(len(A_ ) , A_ ) lowerCamelCase_ , lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = True check_hidden_states_output(A_ , A_ , A_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase_ = True check_hidden_states_output(A_ , A_ , A_ ) def a__ ( self : Any ) -> str: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) def a__ ( self : Any ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A_ ) @slow def a__ ( self : Dict ) -> List[Any]: """simple docstring""" for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ = MobileNetVaModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A( unittest.TestCase ): '''simple docstring''' @cached_property def a__ ( self : List[Any] ) -> Union[str, Any]: """simple docstring""" return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v2_1.0_224' ) if is_vision_available() else None ) @slow def a__ ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" lowerCamelCase_ = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v2_1.0_224' ).to(A_ ) lowerCamelCase_ = self.default_image_processor lowerCamelCase_ = prepare_img() lowerCamelCase_ = image_processor(images=A_ , return_tensors='pt' ).to(A_ ) # forward pass with torch.no_grad(): lowerCamelCase_ = model(**A_ ) # verify the logits lowerCamelCase_ = torch.Size((1, 1001) ) self.assertEqual(outputs.logits.shape , A_ ) lowerCamelCase_ = torch.tensor([0.2445, -1.1993, 0.1905] ).to(A_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , A_ , atol=1E-4 ) ) @slow def a__ ( self : Optional[int] ) -> Dict: """simple docstring""" lowerCamelCase_ = MobileNetVaForSemanticSegmentation.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) lowerCamelCase_ = model.to(A_ ) lowerCamelCase_ = MobileNetVaImageProcessor.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) lowerCamelCase_ = prepare_img() lowerCamelCase_ = image_processor(images=A_ , return_tensors='pt' ).to(A_ ) # forward pass with torch.no_grad(): lowerCamelCase_ = model(**A_ ) lowerCamelCase_ = outputs.logits # verify the logits lowerCamelCase_ = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , A_ ) lowerCamelCase_ = torch.tensor( [ [[17.5790, 17.7581, 18.3355], [18.3257, 18.4230, 18.8973], [18.6169, 18.8650, 19.2187]], [[-2.1595, -2.0977, -2.3741], [-2.4226, -2.3028, -2.6835], [-2.7819, -2.5991, -2.7706]], [[4.2058, 4.8317, 4.7638], [4.4136, 5.0361, 4.9383], [4.5028, 4.9644, 4.8734]], ] , device=A_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , A_ , atol=1E-4 ) )
208
import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings lowerCamelCase : Any = r"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information.\n\n Args:\n title_sep (`str`, *optional*, defaults to `\" / \"`):\n Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`].\n doc_sep (`str`, *optional*, defaults to `\" // \"`):\n Separator inserted between the text of the retrieved document and the original input when calling\n [`RagRetriever`].\n n_docs (`int`, *optional*, defaults to 5):\n Number of documents to retrieve.\n max_combined_length (`int`, *optional*, defaults to 300):\n Max length of contextualized input returned by [`~RagRetriever.__call__`].\n retrieval_vector_size (`int`, *optional*, defaults to 768):\n Dimensionality of the document embeddings indexed by [`RagRetriever`].\n retrieval_batch_size (`int`, *optional*, defaults to 8):\n Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated\n [`RagRetriever`].\n dataset (`str`, *optional*, defaults to `\"wiki_dpr\"`):\n A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids\n using `datasets.list_datasets()`).\n dataset_split (`str`, *optional*, defaults to `\"train\"`)\n Which split of the `dataset` to load.\n index_name (`str`, *optional*, defaults to `\"compressed\"`)\n The index name of the index associated with the `dataset`. One can choose between `\"legacy\"`, `\"exact\"` and\n `\"compressed\"`.\n index_path (`str`, *optional*)\n The path to the serialized faiss index on disk.\n passages_path (`str`, *optional*):\n A path to text passages compatible with the faiss index. Required if using\n [`~models.rag.retrieval_rag.LegacyIndex`]\n use_dummy_dataset (`bool`, *optional*, defaults to `False`)\n Whether to load a \"dummy\" variant of the dataset specified by `dataset`.\n label_smoothing (`float`, *optional*, defaults to 0.0):\n Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing\n in the loss calculation. If set to 0, no label smoothing is performed.\n do_marginalize (`bool`, *optional*, defaults to `False`):\n If `True`, the logits are marginalized over all documents by making use of\n `torch.nn.functional.log_softmax`.\n reduce_loss (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation.\n do_deduplication (`bool`, *optional*, defaults to `True`):\n Whether or not to deduplicate the generations from different context documents for a given input. Has to be\n set to `False` if used while training with distributed backend.\n exclude_bos_score (`bool`, *optional*, defaults to `False`):\n Whether or not to disregard the BOS token when computing the loss.\n output_retrieved(`bool`, *optional*, defaults to `False`):\n If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and\n `context_attention_mask` are returned. See returned tensors for more detail.\n use_cache (`bool`, *optional*, defaults to `True`):\n Whether or not the model should return the last key/values attentions (not used by all models).\n forced_eos_token_id (`int`, *optional*):\n The id of the token to force as the last generated token when `max_length` is reached. Usually set to\n `eos_token_id`.\n" @add_start_docstrings(UpperCamelCase ) class A( UpperCamelCase ): '''simple docstring''' UpperCamelCase = '''rag''' UpperCamelCase = True def __init__( self : Optional[Any] , A_ : Optional[Any]=None , A_ : Any=True , A_ : Dict=None , A_ : Optional[int]=None , A_ : str=None , A_ : int=None , A_ : List[Any]=None , A_ : List[str]=" / " , A_ : Tuple=" // " , A_ : Union[str, Any]=5 , A_ : Optional[Any]=300 , A_ : int=768 , A_ : Dict=8 , A_ : int="wiki_dpr" , A_ : int="train" , A_ : List[str]="compressed" , A_ : Tuple=None , A_ : Optional[Any]=None , A_ : Optional[int]=False , A_ : str=False , A_ : Optional[Any]=0.0 , A_ : Union[str, Any]=True , A_ : List[Any]=False , A_ : Union[str, Any]=False , A_ : Dict=False , A_ : str=True , A_ : List[str]=None , **A_ : Optional[Any] , ) -> str: """simple docstring""" super().__init__( bos_token_id=A_ , pad_token_id=A_ , eos_token_id=A_ , decoder_start_token_id=A_ , forced_eos_token_id=A_ , is_encoder_decoder=A_ , prefix=A_ , vocab_size=A_ , **A_ , ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" lowerCamelCase_ = kwargs.pop('question_encoder' ) lowerCamelCase_ = question_encoder_config.pop('model_type' ) lowerCamelCase_ = kwargs.pop('generator' ) lowerCamelCase_ = decoder_config.pop('model_type' ) from ..auto.configuration_auto import AutoConfig lowerCamelCase_ = AutoConfig.for_model(A_ , **A_ ) lowerCamelCase_ = AutoConfig.for_model(A_ , **A_ ) lowerCamelCase_ = reduce_loss lowerCamelCase_ = label_smoothing lowerCamelCase_ = exclude_bos_score lowerCamelCase_ = do_marginalize lowerCamelCase_ = title_sep lowerCamelCase_ = doc_sep lowerCamelCase_ = n_docs lowerCamelCase_ = max_combined_length lowerCamelCase_ = dataset lowerCamelCase_ = dataset_split lowerCamelCase_ = index_name lowerCamelCase_ = retrieval_vector_size lowerCamelCase_ = retrieval_batch_size lowerCamelCase_ = passages_path lowerCamelCase_ = index_path lowerCamelCase_ = use_dummy_dataset lowerCamelCase_ = output_retrieved lowerCamelCase_ = do_deduplication lowerCamelCase_ = use_cache if self.forced_eos_token_id is None: lowerCamelCase_ = getattr(self.generator , 'forced_eos_token_id' , A_ ) @classmethod def a__ ( cls : str , A_ : PretrainedConfig , A_ : PretrainedConfig , **A_ : List[str] ) -> PretrainedConfig: """simple docstring""" return cls(question_encoder=question_encoder_config.to_dict() , generator=generator_config.to_dict() , **A_ ) def a__ ( self : Tuple ) -> int: """simple docstring""" lowerCamelCase_ = copy.deepcopy(self.__dict__ ) lowerCamelCase_ = self.question_encoder.to_dict() lowerCamelCase_ = self.generator.to_dict() lowerCamelCase_ = self.__class__.model_type return output
208
1
"""simple docstring""" import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer _UpperCamelCase : int = ["gpt2"] _UpperCamelCase : Tuple = "gpt2" if is_tf_available(): class UpperCAmelCase_ ( tf.Module): def __init__( self , a ) -> int: super().__init__() lowercase__ : int = tokenizer lowercase__ : Optional[int] = AutoConfig.from_pretrained(a ) lowercase__ : List[str] = TFGPTaLMHeadModel.from_config(a ) @tf.function(input_signature=(tf.TensorSpec((None,) , tf.string , name='text' ),) ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Union[str, Any] = self.tokenizer(a ) lowercase__ : Tuple = tokenized["input_ids"].to_tensor() lowercase__ : int = tf.cast(input_ids_dense > 0 , tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) lowercase__ : Optional[int] = self.model(input_ids=a , attention_mask=a )["logits"] return outputs @require_tf @require_keras_nlp class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> List[str]: super().setUp() lowercase__ : Union[str, Any] = [GPTaTokenizer.from_pretrained(a ) for checkpoint in (TOKENIZER_CHECKPOINTS)] lowercase__ : Optional[Any] = [TFGPTaTokenizer.from_pretrained(a ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) lowercase__ : Tuple = [ "This is a straightforward English test sentence.", "This one has some weird characters\rto\nsee\r\nif those\u00E9break things.", "Now we're going to add some Chinese: 一 二 三 一二三", "And some much more rare Chinese: 齉 堃 齉堃", "Je vais aussi écrire en français pour tester les accents", "Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ", ] lowercase__ : Optional[int] = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def _UpperCAmelCase ( self ) -> Any: for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in self.test_sentences: lowercase__ : Any = tokenizer([test_inputs] , return_tensors='tf' ) lowercase__ : Optional[Any] = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors lowercase__ : Any = python_outputs[key].numpy() lowercase__ : Tuple = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(a , tf.intaa ) == tf_outputs_values ) ) @slow def _UpperCAmelCase ( self ) -> Union[str, Any]: for tf_tokenizer in self.tf_tokenizers: lowercase__ : Tuple = tf.function(a ) for test_inputs in self.test_sentences: lowercase__ : str = tf.constant(a ) lowercase__ : Dict = compiled_tokenizer(a ) lowercase__ : Optional[int] = tf_tokenizer(a ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def _UpperCAmelCase ( self ) -> str: for tf_tokenizer in self.tf_tokenizers: lowercase__ : Union[str, Any] = ModelToSave(tokenizer=a ) lowercase__ : int = tf.convert_to_tensor([self.test_sentences[0]] ) lowercase__ : Union[str, Any] = model.serving(a ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: lowercase__ : int = Path(a ) / "saved.model" tf.saved_model.save(a , a , signatures={'serving_default': model.serving} ) lowercase__ : Any = tf.saved_model.load(a ) lowercase__ : Dict = loaded_model.signatures["serving_default"](a )["output_0"] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def _UpperCAmelCase ( self ) -> Any: for tf_tokenizer in self.tf_tokenizers: lowercase__ : Dict = tf.convert_to_tensor([self.test_sentences[0]] ) lowercase__ : Any = tf_tokenizer(a ) # Build model with some sample inputs lowercase__ : Any = tf_tokenizer.get_config() lowercase__ : List[Any] = TFGPTaTokenizer.from_config(a ) lowercase__ : List[str] = model_from_config(a ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def _UpperCAmelCase ( self ) -> List[str]: for tf_tokenizer in self.tf_tokenizers: # for the test to run lowercase__ : str = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: lowercase__ : Union[str, Any] = tf.convert_to_tensor([self.test_sentences[0]] ) lowercase__ : Any = tf_tokenizer(a , max_length=a ) lowercase__ : str = out["input_ids"].numpy().shape[1] assert out_length == max_length
77
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
326
0
"""simple docstring""" def a__ ( lowerCAmelCase , lowerCAmelCase ) -> str: if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) UpperCAmelCase__ : Optional[Any] = str(bin(lowerCAmelCase ) )[2:] # remove the leading "0b" UpperCAmelCase__ : str = str(bin(lowerCAmelCase ) )[2:] UpperCAmelCase__ : Any = max(len(lowerCAmelCase ) , len(lowerCAmelCase ) ) return "0b" + "".join( str(int("""1""" in (char_a, char_b) ) ) for char_a, char_b in zip(a_binary.zfill(lowerCAmelCase ) , b_binary.zfill(lowerCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
166
"""simple docstring""" import inspect import unittest import warnings from transformers import DeiTConfig from transformers.models.auto import get_values from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_MAPPING, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, DeiTModel, ) from transformers.models.deit.modeling_deit import DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowerCamelCase : '''simple docstring''' def __init__(self , _lowerCamelCase , _lowerCamelCase=13 , _lowerCamelCase=30 , _lowerCamelCase=2 , _lowerCamelCase=3 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=32 , _lowerCamelCase=5 , _lowerCamelCase=4 , _lowerCamelCase=37 , _lowerCamelCase="gelu" , _lowerCamelCase=0.1 , _lowerCamelCase=0.1 , _lowerCamelCase=10 , _lowerCamelCase=0.02 , _lowerCamelCase=3 , _lowerCamelCase=None , _lowerCamelCase=2 , ): """simple docstring""" UpperCAmelCase__ : Dict = parent UpperCAmelCase__ : str = batch_size UpperCAmelCase__ : Optional[int] = image_size UpperCAmelCase__ : Tuple = patch_size UpperCAmelCase__ : Any = num_channels UpperCAmelCase__ : Union[str, Any] = is_training UpperCAmelCase__ : Optional[int] = use_labels UpperCAmelCase__ : List[str] = hidden_size UpperCAmelCase__ : int = num_hidden_layers UpperCAmelCase__ : List[str] = num_attention_heads UpperCAmelCase__ : Dict = intermediate_size UpperCAmelCase__ : Optional[int] = hidden_act UpperCAmelCase__ : Any = hidden_dropout_prob UpperCAmelCase__ : Optional[Any] = attention_probs_dropout_prob UpperCAmelCase__ : Dict = type_sequence_label_size UpperCAmelCase__ : Optional[int] = initializer_range UpperCAmelCase__ : str = scope UpperCAmelCase__ : Optional[Any] = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) UpperCAmelCase__ : int = (image_size // patch_size) ** 2 UpperCAmelCase__ : Tuple = num_patches + 2 def _a (self ): """simple docstring""" UpperCAmelCase__ : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ : Union[str, Any] = None if self.use_labels: UpperCAmelCase__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : int = self.get_config() return config, pixel_values, labels def _a (self ): """simple docstring""" return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCamelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Tuple = DeiTModel(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Union[str, Any] = model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Optional[int] = DeiTForMaskedImageModeling(config=_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Optional[int] = model(_lowerCamelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images UpperCAmelCase__ : str = 1 UpperCAmelCase__ : List[str] = DeiTForMaskedImageModeling(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ : Dict = model(_lowerCamelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): """simple docstring""" UpperCAmelCase__ : Optional[Any] = self.type_sequence_label_size UpperCAmelCase__ : List[str] = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : str = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images UpperCAmelCase__ : Union[str, Any] = 1 UpperCAmelCase__ : int = DeiTForImageClassification(_lowerCamelCase ) model.to(_lowerCamelCase ) model.eval() UpperCAmelCase__ : Optional[int] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCAmelCase__ : Union[str, Any] = model(_lowerCamelCase , labels=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _a (self ): """simple docstring""" UpperCAmelCase__ : List[Any] = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : Tuple = config_and_inputs UpperCAmelCase__ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class lowerCamelCase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = ( ( DeiTModel, DeiTForImageClassification, DeiTForImageClassificationWithTeacher, DeiTForMaskedImageModeling, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE = ( { 'feature-extraction': DeiTModel, 'image-classification': (DeiTForImageClassification, DeiTForImageClassificationWithTeacher), } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = False def _a (self ): """simple docstring""" UpperCAmelCase__ : Optional[int] = DeiTModelTester(self ) UpperCAmelCase__ : Optional[Any] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase , hidden_size=37 ) def _a (self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="""DeiT does not use inputs_embeds""" ) def _a (self ): """simple docstring""" pass def _a (self ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Optional[Any] = model_class(_lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCAmelCase__ : Optional[int] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCamelCase , nn.Linear ) ) def _a (self ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ : Any = model_class(_lowerCamelCase ) UpperCAmelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ : Union[str, Any] = [*signature.parameters.keys()] UpperCAmelCase__ : str = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCamelCase ) def _a (self ): """simple docstring""" UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) def _a (self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=False ): """simple docstring""" UpperCAmelCase__ : Optional[int] = super()._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if return_labels: if model_class.__name__ == "DeiTForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def _a (self ): """simple docstring""" if not self.model_tester.is_training: return UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : List[str] = True for model_class in self.all_model_classes: # DeiTForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(_lowerCamelCase ) or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue UpperCAmelCase__ : Dict = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() UpperCAmelCase__ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) UpperCAmelCase__ : int = model(**_lowerCamelCase ).loss loss.backward() def _a (self ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return UpperCAmelCase__ : List[str] = False UpperCAmelCase__ : Optional[Any] = True for model_class in self.all_model_classes: if model_class in get_values(_lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # DeiTForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "DeiTForImageClassificationWithTeacher": continue UpperCAmelCase__ : Optional[Any] = model_class(_lowerCamelCase ) model.gradient_checkpointing_enable() model.to(_lowerCamelCase ) model.train() UpperCAmelCase__ : Union[str, Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) UpperCAmelCase__ : Tuple = model(**_lowerCamelCase ).loss loss.backward() def _a (self ): """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : Optional[Any] = [ {"""title""": """multi_label_classification""", """num_labels""": 2, """dtype""": torch.float}, {"""title""": """single_label_classification""", """num_labels""": 1, """dtype""": torch.long}, {"""title""": """regression""", """num_labels""": 1, """dtype""": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(_lowerCamelCase ), *get_values(_lowerCamelCase ), ] or model_class.__name__ == "DeiTForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"""Testing {model_class} with {problem_type['title']}""" ): UpperCAmelCase__ : List[str] = problem_type["""title"""] UpperCAmelCase__ : List[Any] = problem_type["""num_labels"""] UpperCAmelCase__ : Optional[Any] = model_class(_lowerCamelCase ) model.to(_lowerCamelCase ) model.train() UpperCAmelCase__ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) if problem_type["num_labels"] > 1: UpperCAmelCase__ : Optional[int] = inputs["""labels"""].unsqueeze(1 ).repeat(1 , problem_type["""num_labels"""] ) UpperCAmelCase__ : str = inputs["""labels"""].to(problem_type["""dtype"""] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=_lowerCamelCase ) as warning_list: UpperCAmelCase__ : Any = model(**_lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def _a (self ): """simple docstring""" for model_name in DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : int = DeiTModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def a__ ( ) -> int: UpperCAmelCase__ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def _a (self ): """simple docstring""" return ( DeiTImageProcessor.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ) if is_vision_available() else None ) @slow def _a (self ): """simple docstring""" UpperCAmelCase__ : int = DeiTForImageClassificationWithTeacher.from_pretrained("""facebook/deit-base-distilled-patch16-224""" ).to( _lowerCamelCase ) UpperCAmelCase__ : Union[str, Any] = self.default_image_processor UpperCAmelCase__ : Tuple = prepare_img() UpperCAmelCase__ : Tuple = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ).to(_lowerCamelCase ) # forward pass with torch.no_grad(): UpperCAmelCase__ : Any = model(**_lowerCamelCase ) # verify the logits UpperCAmelCase__ : List[str] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) UpperCAmelCase__ : Dict = torch.tensor([-1.0_266, 0.1_912, -1.2_861] ).to(_lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 ) ) @slow @require_accelerate @require_torch_gpu def _a (self ): """simple docstring""" UpperCAmelCase__ : List[str] = DeiTModel.from_pretrained( """facebook/deit-base-distilled-patch16-224""" , torch_dtype=torch.floataa , device_map="""auto""" ) UpperCAmelCase__ : Union[str, Any] = self.default_image_processor UpperCAmelCase__ : int = prepare_img() UpperCAmelCase__ : str = image_processor(images=_lowerCamelCase , return_tensors="""pt""" ) UpperCAmelCase__ : Dict = inputs.pixel_values.to(_lowerCamelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): UpperCAmelCase__ : int = model(_lowerCamelCase )
166
1
from __future__ import annotations from fractions import Fraction def __snake_case ( _UpperCAmelCase , _UpperCAmelCase ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def __snake_case ( _UpperCAmelCase ): __a = [] __a = 11 __a = int('''1''' + '''0''' * digit_len ) for num in range(_UpperCAmelCase , _UpperCAmelCase ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(_UpperCAmelCase , _UpperCAmelCase ): solutions.append(f'{num}/{den}' ) den += 1 num += 1 __a = 10 return solutions def __snake_case ( _UpperCAmelCase = 2 ): __a = 1.0 for fraction in fraction_list(_UpperCAmelCase ): __a = Fraction(_UpperCAmelCase ) result *= frac.denominator / frac.numerator return int(_UpperCAmelCase ) if __name__ == "__main__": print(solution())
49
"""simple docstring""" from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class __lowerCAmelCase : '''simple docstring''' def __init__( self , _a , ): __a = parent __a = 13 __a = 7 __a = True __a = True __a = True __a = 99 __a = 32 __a = 2 __a = 4 __a = 37 __a = '''gelu''' __a = 0.1 __a = 0.1 __a = 512 __a = 16 __a = 2 __a = 0.02 __a = 3 __a = 4 __a = None def __UpperCAmelCase ( self ): __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = None if self.use_input_mask: __a = random_attention_mask([self.batch_size, self.seq_length] ) __a = None __a = None __a = None if self.use_labels: __a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __a = ids_tensor([self.batch_size] , self.num_choices ) __a = EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __UpperCAmelCase ( self ): ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = self.prepare_config_and_inputs() __a = True __a = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __a = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a ): __a = TFEsmModel(config=_a ) __a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} __a = model(_a ) __a = [input_ids, input_mask] __a = model(_a ) __a = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a , _a , _a , ): __a = True __a = TFEsmModel(config=_a ) __a = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''encoder_hidden_states''': encoder_hidden_states, '''encoder_attention_mask''': encoder_attention_mask, } __a = model(_a ) __a = [input_ids, input_mask] __a = model(_a , encoder_hidden_states=_a ) # Also check the case where encoder outputs are not passed __a = model(_a , attention_mask=_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a ): __a = TFEsmForMaskedLM(config=_a ) __a = model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __UpperCAmelCase ( self , _a , _a , _a , _a , _a , _a ): __a = self.num_labels __a = TFEsmForTokenClassification(config=_a ) __a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} __a = model(_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __UpperCAmelCase ( self ): __a = self.prepare_config_and_inputs() ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = config_and_inputs __a = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : int = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) __UpperCAmelCase : Tuple = ( { 'feature-extraction': TFEsmModel, 'fill-mask': TFEsmForMaskedLM, 'text-classification': TFEsmForSequenceClassification, 'token-classification': TFEsmForTokenClassification, 'zero-shot': TFEsmForSequenceClassification, } if is_tf_available() else {} ) __UpperCAmelCase : Tuple = False __UpperCAmelCase : Union[str, Any] = False def __UpperCAmelCase ( self ): __a = TFEsmModelTester(self ) __a = ConfigTester(self , config_class=_a , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_a ) @slow def __UpperCAmelCase ( self ): for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __a = TFEsmModel.from_pretrained(_a ) self.assertIsNotNone(_a ) @unittest.skip('''Protein models do not support embedding resizing.''' ) def __UpperCAmelCase ( self ): pass @unittest.skip('''Protein models do not support embedding resizing.''' ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): __a , __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(_a ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer __a = model.get_bias() assert isinstance(_a , _a ) for k, v in name.items(): assert isinstance(_a , tf.Variable ) else: __a = model.get_output_embeddings() assert x is None __a = model.get_bias() assert name is None @require_tf class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCAmelCase ( self ): __a = TFEsmForMaskedLM.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) __a = tf.constant([[0, 1, 2, 3, 4, 5]] ) __a = model(_a )[0] __a = [1, 6, 33] self.assertEqual(list(output.numpy().shape ) , _a ) # compare the actual values for a slice. __a = tf.constant( [ [ [8.92_1518, -10.58_9814, -6.467_1307], [-6.396_7156, -13.91_1377, -1.121_1915], [-7.78_1247, -13.95_1557, -3.74_0592], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-2 ) ) @slow def __UpperCAmelCase ( self ): __a = TFEsmModel.from_pretrained('''facebook/esm2_t6_8M_UR50D''' ) __a = tf.constant([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) __a = model(_a )[0] # compare the actual values for a slice. __a = tf.constant( [ [ [0.1444_3092, 0.5412_5327, 0.324_7739], [0.3034_0484, 0.0052_6676, 0.3107_7722], [0.3227_8043, -0.2498_7096, 0.341_4628], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
45
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCAmelCase :Any = logging.get_logger(__name__) _lowerCAmelCase :int = { 'facebook/xlm-roberta-xl': 'https://huggingface.co/facebook/xlm-roberta-xl/resolve/main/config.json', 'facebook/xlm-roberta-xxl': 'https://huggingface.co/facebook/xlm-roberta-xxl/resolve/main/config.json', # See all XLM-RoBERTa-XL models at https://huggingface.co/models?filter=xlm-roberta-xl } class _UpperCAmelCase ( _lowercase ): '''simple docstring''' a__ ='''xlm-roberta-xl''' def __init__( self , A=2_5_0_8_8_0 , A=2_5_6_0 , A=3_6 , A=3_2 , A=1_0_2_4_0 , A="gelu" , A=0.1 , A=0.1 , A=5_1_4 , A=1 , A=0.02 , A=1E-05 , A=1 , A=0 , A=2 , A="absolute" , A=True , A=None , **A , ) -> Union[str, Any]: super().__init__(pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) _UpperCAmelCase : Optional[Any] = vocab_size _UpperCAmelCase : Tuple = hidden_size _UpperCAmelCase : str = num_hidden_layers _UpperCAmelCase : Tuple = num_attention_heads _UpperCAmelCase : Tuple = hidden_act _UpperCAmelCase : Optional[int] = intermediate_size _UpperCAmelCase : Optional[Any] = hidden_dropout_prob _UpperCAmelCase : Tuple = attention_probs_dropout_prob _UpperCAmelCase : Optional[int] = max_position_embeddings _UpperCAmelCase : Dict = type_vocab_size _UpperCAmelCase : Tuple = initializer_range _UpperCAmelCase : List[Any] = layer_norm_eps _UpperCAmelCase : str = position_embedding_type _UpperCAmelCase : Dict = use_cache _UpperCAmelCase : Optional[Any] = classifier_dropout class _UpperCAmelCase ( _lowercase ): '''simple docstring''' @property def __lowerCAmelCase ( self ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": _UpperCAmelCase : Dict = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _UpperCAmelCase : Union[str, Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
360
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def lowerCamelCase_ (): _UpperCAmelCase : Optional[int] = [randint(-1000 , 1000 ) for i in range(10 )] _UpperCAmelCase : int = randint(-5000 , 5000 ) return (arr, r) _lowerCAmelCase :Optional[Any] = make_dataset() def lowerCamelCase_ (UpperCamelCase__ : list[int] , UpperCamelCase__ : int ): for triplet in permutations(UpperCamelCase__ , 3 ): if sum(UpperCamelCase__ ) == target: return tuple(sorted(UpperCamelCase__ ) ) return (0, 0, 0) def lowerCamelCase_ (UpperCamelCase__ : list[int] , UpperCamelCase__ : int ): arr.sort() _UpperCAmelCase : Optional[int] = len(UpperCamelCase__ ) for i in range(n - 1 ): _UpperCAmelCase , _UpperCAmelCase : Any = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def lowerCamelCase_ (): _UpperCAmelCase : Union[str, Any] = ''' from __main__ import dataset, triplet_sum1, triplet_sum2 ''' _UpperCAmelCase : Any = ''' triplet_sum1(*dataset) ''' _UpperCAmelCase : List[Any] = ''' triplet_sum2(*dataset) ''' _UpperCAmelCase : List[Any] = repeat(setup=UpperCamelCase__ , stmt=UpperCamelCase__ , repeat=5 , number=1_0000 ) _UpperCAmelCase : List[Any] = repeat(setup=UpperCamelCase__ , stmt=UpperCamelCase__ , repeat=5 , number=1_0000 ) return (min(UpperCamelCase__ ), min(UpperCamelCase__ )) if __name__ == "__main__": from doctest import testmod testmod() _lowerCAmelCase :List[str] = solution_times() print(f"The time for naive implementation is {times[0]}.") print(f"The time for optimized implementation is {times[1]}.")
68
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore _UpperCAmelCase : Any = """ Human: <<task>> Assistant: """ _UpperCAmelCase : List[str] = """huggingface-tools/default-prompts""" _UpperCAmelCase : str = {"""chat""": """chat_prompt_template.txt""", """run""": """run_prompt_template.txt"""} def A ( lowercase , lowercase , lowercase="run" ) -> Dict: '''simple docstring''' if prompt_or_repo_id is None: UpperCamelCase = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search('\\s' , A__ ) is not None: return prompt_or_repo_id UpperCamelCase = cached_file( A__ , PROMPT_FILES[mode] , repo_type='dataset' , user_agent={'agent': agent_name} ) with open(A__ , 'r' , encoding='utf-8' ) as f: return f.read()
222
import argparse import os from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_task_guides.py lowercase : List[str] = """src/transformers""" lowercase : Optional[int] = """docs/source/en/tasks""" def A_ ( A__ , A__ , A__ ) -> Tuple: with open(A__ , 'r' , encoding='utf-8' , newline='\n' ) as f: a__ : Any = f.readlines() # Find the start prompt. a__ : str = 0 while not lines[start_index].startswith(A__ ): start_index += 1 start_index += 1 a__ : int = start_index while not lines[end_index].startswith(A__ ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # This is to make sure the transformers module imported is the one in the repo. lowercase : Tuple = direct_transformers_import(TRANSFORMERS_PATH) lowercase : Optional[Any] = { """asr.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_CTC_MAPPING_NAMES, """audio_classification.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES, """language_modeling.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_CAUSAL_LM_MAPPING_NAMES, """image_classification.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES, """masked_language_modeling.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_MASKED_LM_MAPPING_NAMES, """multiple_choice.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES, """object_detection.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_OBJECT_DETECTION_MAPPING_NAMES, """question_answering.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES, """semantic_segmentation.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEMANTIC_SEGMENTATION_MAPPING_NAMES, """sequence_classification.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES, """summarization.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, """token_classification.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES, """translation.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES, """video_classification.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING_NAMES, """document_question_answering.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING_NAMES, """monocular_depth_estimation.md""": transformers_module.models.auto.modeling_auto.MODEL_FOR_DEPTH_ESTIMATION_MAPPING_NAMES, } # This list contains model types used in some task guides that are not in `CONFIG_MAPPING_NAMES` (therefore not in any # `MODEL_MAPPING_NAMES` or any `MODEL_FOR_XXX_MAPPING_NAMES`). lowercase : Optional[Any] = { """summarization.md""": ("""nllb""",), """translation.md""": ("""nllb""",), } def A_ ( A__ ) -> Optional[int]: a__ : Optional[Any] = TASK_GUIDE_TO_MODELS[task_guide] a__ : int = SPECIAL_TASK_GUIDE_TO_MODEL_TYPES.get(A__ , set() ) a__ : Optional[Any] = { code: name for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if (code in model_maping_names or code in special_model_types) } return ", ".join([F'[{name}](../model_doc/{code})' for code, name in model_names.items()] ) + "\n" def A_ ( A__ , A__=False ) -> Optional[int]: a__ , a__ , a__ , a__ : Dict = _find_text_in_file( filename=os.path.join(A__ , A__ ) , start_prompt='<!--This tip is automatically generated by `make fix-copies`, do not fill manually!-->' , end_prompt='<!--End of the generated tip-->' , ) a__ : List[Any] = get_model_list_for_task(A__ ) if current_list != new_list: if overwrite: with open(os.path.join(A__ , A__ ) , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(lines[:start_index] + [new_list] + lines[end_index:] ) else: raise ValueError( F'The list of models that can be used in the {task_guide} guide needs an update. Run `make fix-copies`' ' to fix this.' ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") lowercase : str = parser.parse_args() for task_guide in TASK_GUIDE_TO_MODELS.keys(): check_model_list_for_task(task_guide, args.fix_and_overwrite)
99
0
from math import pow def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , ): if current_sum == needed_sum: # If the sum of the powers is equal to needed_sum, then we have a solution. solutions_count += 1 return current_sum, solutions_count A : Tuple = int(pow(lowerCamelCase_ , lowerCamelCase_ ) ) if current_sum + i_to_n <= needed_sum: # If the sum of the powers is less than needed_sum, then continue adding powers. current_sum += i_to_n A : int = backtrack( lowerCamelCase_ , lowerCamelCase_ , current_number + 1 , lowerCamelCase_ , lowerCamelCase_ ) current_sum -= i_to_n if i_to_n < needed_sum: # If the power of i is less than needed_sum, then try with the next power. A : Any = backtrack( lowerCamelCase_ , lowerCamelCase_ , current_number + 1 , lowerCamelCase_ , lowerCamelCase_ ) return current_sum, solutions_count def UpperCAmelCase ( _lowerCamelCase , _lowerCamelCase ): if not (1 <= needed_sum <= 1000 and 2 <= power <= 10): raise ValueError( "Invalid input\n" "needed_sum must be between 1 and 1000, power between 2 and 10." ) return backtrack(lowerCamelCase_ , lowerCamelCase_ , 1 , 0 , 0 )[1] # Return the solutions_count if __name__ == "__main__": import doctest doctest.testmod()
356
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE = {"""configuration_ibert""": ["""IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """IBertConfig""", """IBertOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE = [ """IBERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """IBertForMaskedLM""", """IBertForMultipleChoice""", """IBertForQuestionAnswering""", """IBertForSequenceClassification""", """IBertForTokenClassification""", """IBertModel""", """IBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
256
0
"""simple docstring""" from manim import * class snake_case__ ( snake_case_ ): def a__ ( self ): __a = Rectangle(height=0.5 , width=0.5 ) __a = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) __a = Rectangle(height=0.25 , width=0.25 ) __a = [mem.copy() for i in range(6 )] __a = [mem.copy() for i in range(6 )] __a = VGroup(*lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = VGroup(*lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = VGroup(lowerCamelCase , lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = Text("CPU" , font_size=24 ) __a = Group(lowerCamelCase , lowerCamelCase ).arrange(lowerCamelCase , buff=0.5 , aligned_edge=lowerCamelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(lowerCamelCase ) __a = [mem.copy() for i in range(4 )] __a = VGroup(*lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = Text("GPU" , font_size=24 ) __a = Group(lowerCamelCase , lowerCamelCase ).arrange(lowerCamelCase , buff=0.5 , aligned_edge=lowerCamelCase ) gpu.move_to([-1, -1, 0] ) self.add(lowerCamelCase ) __a = [mem.copy() for i in range(6 )] __a = VGroup(*lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = Text("Model" , font_size=24 ) __a = Group(lowerCamelCase , lowerCamelCase ).arrange(lowerCamelCase , buff=0.5 , aligned_edge=lowerCamelCase ) model.move_to([3, -1.0, 0] ) self.add(lowerCamelCase ) __a = [] __a = [] for i, rect in enumerate(lowerCamelCase ): __a = fill.copy().set_fill(lowerCamelCase , opacity=0.8 ) target.move_to(lowerCamelCase ) model_arr.append(lowerCamelCase ) __a = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(lowerCamelCase , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(lowerCamelCase ) self.add(*lowerCamelCase , *lowerCamelCase ) __a = [meta_mem.copy() for i in range(6 )] __a = [meta_mem.copy() for i in range(6 )] __a = VGroup(*lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = VGroup(*lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = VGroup(lowerCamelCase , lowerCamelCase ).arrange(lowerCamelCase , buff=0 ) __a = Text("Disk" , font_size=24 ) __a = Group(lowerCamelCase , lowerCamelCase ).arrange(lowerCamelCase , buff=0.5 , aligned_edge=lowerCamelCase ) disk.move_to([-4, -1.25, 0] ) self.add(lowerCamelCase , lowerCamelCase ) __a = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) __a = MarkupText( F"<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(lowerCamelCase , lowerCamelCase ) __a = MarkupText( F"<span fgcolor='{BLUE}'>●</span> Checkpoint" , font_size=18 , ) blue_text.next_to(lowerCamelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(lowerCamelCase ) __a = MarkupText( F"Now watch as an input is passed through the model\nand how the memory is utilized and handled." , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase ) ) __a = Square(0.3 ) input.set_fill(lowerCamelCase , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , lowerCamelCase , buff=0.5 ) self.play(Write(lowerCamelCase ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=lowerCamelCase , buff=0.02 ) self.play(MoveToTarget(lowerCamelCase ) ) self.play(FadeOut(lowerCamelCase ) ) __a = Arrow(start=lowerCamelCase , end=lowerCamelCase , color=lowerCamelCase , buff=0.5 ) a.next_to(model_arr[0].get_left() , lowerCamelCase , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) __a = MarkupText( F"As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back." , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase , run_time=3 ) ) __a = {"run_time": 1, "fade_in": True, "fade_out": True, "buff": 0.02} self.play( Write(lowerCamelCase ) , Circumscribe(model_arr[0] , color=lowerCamelCase , **lowerCamelCase ) , Circumscribe(model_cpu_arr[0] , color=lowerCamelCase , **lowerCamelCase ) , Circumscribe(gpu_rect[0] , color=lowerCamelCase , **lowerCamelCase ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) __a = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , lowerCamelCase , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) __a = AnimationGroup( FadeOut(lowerCamelCase , run_time=0.5 ) , MoveToTarget(lowerCamelCase , run_time=0.5 ) , FadeIn(lowerCamelCase , run_time=0.5 ) , lag_ratio=0.2 ) self.play(lowerCamelCase ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: __a = 0.7 self.play( Circumscribe(model_arr[i] , **lowerCamelCase ) , Circumscribe(cpu_left_col_base[i] , **lowerCamelCase ) , Circumscribe(cpu_left_col_base[i + 1] , color=lowerCamelCase , **lowerCamelCase ) , Circumscribe(gpu_rect[0] , color=lowerCamelCase , **lowerCamelCase ) , Circumscribe(model_arr[i + 1] , color=lowerCamelCase , **lowerCamelCase ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=lowerCamelCase , **lowerCamelCase ) , Circumscribe(cpu_left_col_base[-1] , color=lowerCamelCase , **lowerCamelCase ) , Circumscribe(gpu_rect[0] , color=lowerCamelCase , **lowerCamelCase ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) __a = a_c __a = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(lowerCamelCase ) , FadeOut(lowerCamelCase , run_time=0.5 ) , ) __a = MarkupText(F"Inference on a model too large for GPU memory\nis successfully completed." , font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(lowerCamelCase , run_time=3 ) , MoveToTarget(lowerCamelCase ) ) self.wait()
261
"""simple docstring""" import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class snake_case__ : def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=99 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=37 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=50 , lowerCamelCase=0.02 , lowerCamelCase=True , lowerCamelCase=None , ): __a = parent __a = batch_size __a = seq_length __a = is_training __a = use_input_mask __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = initializer_range __a = use_labels __a = scope def a__ ( self ): __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = None if self.use_input_mask: __a = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: __a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __a = self.get_config() return config, input_ids, input_mask, token_labels def a__ ( self ): return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , ) def a__ ( self ): ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = self.prepare_config_and_inputs() __a = True __a = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __a = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): __a = BertGenerationEncoder(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , attention_mask=lowerCamelCase ) __a = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): __a = True __a = BertGenerationEncoder(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , ) __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ): __a = True __a = True __a = BertGenerationDecoder(config=lowerCamelCase ).to(lowerCamelCase ).eval() # first forward pass __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , use_cache=lowerCamelCase , ) __a = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __a = ids_tensor((self.batch_size, 3) , config.vocab_size ) __a = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __a = torch.cat([input_ids, next_tokens] , dim=-1 ) __a = torch.cat([input_mask, next_mask] , dim=-1 ) __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] __a = model( lowerCamelCase , attention_mask=lowerCamelCase , encoder_hidden_states=lowerCamelCase , encoder_attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] # select random slice __a = ids_tensor((1,) , output_from_past.shape[-1] ).item() __a = output_from_no_past[:, -3:, random_slice_idx].detach() __a = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , *lowerCamelCase , ): __a = BertGenerationDecoder(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __a = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__ ( self ): __a , __a , __a , __a = self.prepare_config_and_inputs() __a = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class snake_case__ ( snake_case_, snake_case_, snake_case_, unittest.TestCase ): _snake_case : Union[str, Any] = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () _snake_case : Any = (BertGenerationDecoder,) if is_torch_available() else () _snake_case : Union[str, Any] = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def a__ ( self ): __a = BertGenerationEncoderTester(self ) __a = ConfigTester(self , config_class=lowerCamelCase , hidden_size=37 ) def a__ ( self ): self.config_tester.run_common_tests() def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def a__ ( self ): __a , __a , __a , __a = self.model_tester.prepare_config_and_inputs() __a = "bert" self.model_tester.create_and_check_model(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*lowerCamelCase ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*lowerCamelCase ) def a__ ( self ): # This regression test was failing with PyTorch < 1.3 ( ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ( __a ) , ) = self.model_tester.prepare_config_and_inputs_for_decoder() __a = None self.model_tester.create_and_check_model_as_decoder( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) def a__ ( self ): __a = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*lowerCamelCase ) @slow def a__ ( self ): __a = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) self.assertIsNotNone(lowerCamelCase ) @require_torch class snake_case__ ( unittest.TestCase ): @slow def a__ ( self ): __a = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) __a = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 10140, 102]] ) with torch.no_grad(): __a = model(lowerCamelCase )[0] __a = torch.Size([1, 8, 1024] ) self.assertEqual(output.shape , lowerCamelCase ) __a = torch.tensor( [[[0.1775, 0.0083, -0.0321], [1.6002, 0.1287, 0.3912], [2.1473, 0.5791, 0.6066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @require_torch class snake_case__ ( unittest.TestCase ): @slow def a__ ( self ): __a = BertGenerationDecoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) __a = torch.tensor([[101, 7592, 1010, 2026, 3899, 2003, 10140, 102]] ) with torch.no_grad(): __a = model(lowerCamelCase )[0] __a = torch.Size([1, 8, 50358] ) self.assertEqual(output.shape , lowerCamelCase ) __a = torch.tensor( [[[-0.5788, -2.5994, -3.7054], [0.0438, 4.7997, 1.8795], [1.5862, 6.6409, 4.4638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase , atol=1E-4 ) )
261
1
from __future__ import annotations from PIL import Image # Define glider example _snake_case = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example _snake_case = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def _A ( __magic_name__ ): lowercase__ = [] for i in range(len(lowerCAmelCase__ ) ): lowercase__ = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours lowercase__ = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(lowerCAmelCase__ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(lowerCAmelCase__ ) - 1: neighbour_count += cells[i + 1][j] if i < len(lowerCAmelCase__ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. lowercase__ = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(lowerCAmelCase__ ) return next_generation def _A ( __magic_name__ , __magic_name__ ): lowercase__ = [] for _ in range(lowerCAmelCase__ ): # Create output image lowercase__ = Image.new("RGB" , (len(cells[0] ), len(lowerCAmelCase__ )) ) lowercase__ = img.load() # Save cells to image for x in range(len(lowerCAmelCase__ ) ): for y in range(len(cells[0] ) ): lowercase__ = 255 - cells[y][x] * 255 lowercase__ = (colour, colour, colour) # Save image images.append(lowerCAmelCase__ ) lowercase__ = new_generation(lowerCAmelCase__ ) return images if __name__ == "__main__": _snake_case = generate_images(GLIDER, 16) images[0].save("""out.gif""", save_all=True, append_images=images[1:])
366
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { """configuration_autoformer""": [ """AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """AutoformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """AutoformerForPrediction""", """AutoformerModel""", """AutoformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
201
0
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. UpperCAmelCase : Tuple = {"""LayoutLMv2Config""", """LayoutLMv3Config"""} @is_pipeline_test class __lowerCAmelCase ( unittest.TestCase): _lowercase : Dict = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING _lowercase : int = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: _lowercase : Dict = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: _lowercase : Optional[Any] = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } @require_torch def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ : Union[str, Any] =pipeline( task="text-classification" , model="hf-internal-testing/tiny-random-distilbert" , framework="pt" ) a__ : str =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) a__ : Tuple =text_classifier("This is great !" , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}] ) a__ : Optional[int] =text_classifier(["This is great !", "This is bad"] , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], ] , ) a__ : Any =text_classifier("This is great !" , top_k=1 ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) # Legacy behavior a__ : Optional[int] =text_classifier("This is great !" , return_all_scores=lowerCAmelCase__ ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) a__ : Any =text_classifier("This is great !" , return_all_scores=lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [[{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}]] ) a__ : int =text_classifier(["This is great !", "Something else"] , return_all_scores=lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], [{"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_1", "score": 0.4_96}], ] , ) a__ : List[Any] =text_classifier(["This is great !", "Something else"] , return_all_scores=lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [ {"label": "LABEL_0", "score": 0.5_04}, {"label": "LABEL_0", "score": 0.5_04}, ] , ) @require_torch def _lowercase ( self ) -> Dict: '''simple docstring''' import torch a__ : Any =pipeline( task="text-classification" , model="hf-internal-testing/tiny-random-distilbert" , framework="pt" , device=torch.device("cpu" ) , ) a__ : Optional[int] =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) @require_tf def _lowercase ( self ) -> Any: '''simple docstring''' a__ : Optional[Any] =pipeline( task="text-classification" , model="hf-internal-testing/tiny-random-distilbert" , framework="tf" ) a__ : List[str] =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "LABEL_0", "score": 0.5_04}] ) @slow @require_torch def _lowercase ( self ) -> str: '''simple docstring''' a__ : Optional[Any] =pipeline("text-classification" ) a__ : Dict =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 1.0}] ) a__ : Tuple =text_classifier("This is bad !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "NEGATIVE", "score": 1.0}] ) a__ : Union[str, Any] =text_classifier("Birds are a type of animal" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 0.9_88}] ) @slow @require_tf def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ : Optional[int] =pipeline("text-classification" , framework="tf" ) a__ : str =text_classifier("This is great !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 1.0}] ) a__ : List[str] =text_classifier("This is bad !" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "NEGATIVE", "score": 1.0}] ) a__ : List[str] =text_classifier("Birds are a type of animal" ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": "POSITIVE", "score": 0.9_88}] ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[Any]: '''simple docstring''' a__ : Optional[int] =TextClassificationPipeline(model=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) return text_classifier, ["HuggingFace is in", "This is another test"] def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> List[str]: '''simple docstring''' a__ : int =text_classifier.model # Small inputs because BartTokenizer tiny has maximum position embeddings = 22 a__ : Optional[Any] ="HuggingFace is in" a__ : List[str] =text_classifier(lowerCAmelCase__ ) self.assertEqual(nested_simplify(lowerCAmelCase__ ) , [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] ) self.assertTrue(outputs[0]["label"] in model.config.idalabel.values() ) a__ : Optional[int] =["HuggingFace is in ", "Paris is in France"] a__ : Optional[Any] =text_classifier(lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}, {"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] , ) self.assertTrue(outputs[0]["label"] in model.config.idalabel.values() ) self.assertTrue(outputs[1]["label"] in model.config.idalabel.values() ) # Forcing to get all results with `top_k=None` # This is NOT the legacy format a__ : List[str] =text_classifier(lowerCAmelCase__ , top_k=lowerCAmelCase__ ) a__ : Any =len(model.config.idalabel.values() ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [[{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] * N, [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] * N] , ) a__ : Optional[int] ={"text": "HuggingFace is in ", "text_pair": "Paris is in France"} a__ : str =text_classifier(lowerCAmelCase__ ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , {"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )} , ) self.assertTrue(outputs["label"] in model.config.idalabel.values() ) # This might be used a text pair, but tokenizer + pipe interaction # makes it hard to understand that it's not using the pair properly # https://github.com/huggingface/transformers/issues/17305 # We disabled this usage instead as it was outputting wrong outputs. a__ : List[str] =[["HuggingFace is in ", "Paris is in France"]] with self.assertRaises(lowerCAmelCase__ ): text_classifier(lowerCAmelCase__ ) # This used to be valid for doing text pairs # We're keeping it working because of backward compatibility a__ : Dict =text_classifier([[["HuggingFace is in ", "Paris is in France"]]] ) self.assertEqual( nested_simplify(lowerCAmelCase__ ) , [{"label": ANY(lowerCAmelCase__ ), "score": ANY(lowerCAmelCase__ )}] , ) self.assertTrue(outputs[0]["label"] in model.config.idalabel.values() )
95
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { '''google/mobilenet_v2_1.4_224''': '''https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json''', '''google/mobilenet_v2_1.0_224''': '''https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json''', '''google/mobilenet_v2_0.75_160''': '''https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json''', '''google/mobilenet_v2_0.35_96''': '''https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json''', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' __A : Dict = "mobilenet_v2" def __init__( self , __A=3 , __A=224 , __A=1.0 , __A=8 , __A=8 , __A=6 , __A=32 , __A=True , __A=True , __A="relu6" , __A=True , __A=0.8 , __A=0.02 , __A=0.001 , __A=255 , **__A , ): """simple docstring""" super().__init__(**__A ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) lowerCamelCase : str = num_channels lowerCamelCase : Any = image_size lowerCamelCase : Union[str, Any] = depth_multiplier lowerCamelCase : Tuple = depth_divisible_by lowerCamelCase : Dict = min_depth lowerCamelCase : Dict = expand_ratio lowerCamelCase : Optional[Any] = output_stride lowerCamelCase : int = first_layer_is_expansion lowerCamelCase : Union[str, Any] = finegrained_output lowerCamelCase : Optional[Any] = hidden_act lowerCamelCase : Optional[Any] = tf_padding lowerCamelCase : Optional[Any] = classifier_dropout_prob lowerCamelCase : Dict = initializer_range lowerCamelCase : str = layer_norm_eps lowerCamelCase : Optional[Any] = semantic_loss_ignore_index class UpperCAmelCase_ ( UpperCamelCase ): '''simple docstring''' __A : Union[str, Any] = version.parse("1.11" ) @property def _snake_case ( self ): """simple docstring""" return OrderedDict([("pixel_values", {0: "batch"})] ) @property def _snake_case ( self ): """simple docstring""" if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def _snake_case ( self ): """simple docstring""" return 1e-4
283
0
'''simple docstring''' import math def lowerCamelCase__ ( _A ): if not isinstance(_A , _A ): a : Union[str, Any] = f"""Input value of [number={number}] must be an integer""" raise TypeError(_A ) if number < 1: a : Dict = f"""Input value of [number={number}] must be > 0""" raise ValueError(_A ) elif number == 1: return 3 elif number == 2: return 5 else: a : Tuple = int(math.log(number // 3 , 2 ) ) + 2 a : Union[str, Any] = [3, 5] a : Union[str, Any] = 2 a : int = 3 for block in range(1 , _A ): for _ in range(_A ): proth_list.append(2 ** (block + 1) + proth_list[proth_index - 1] ) proth_index += 1 increment *= 2 return proth_list[number - 1] if __name__ == "__main__": import doctest doctest.testmod() for number in range(1_1): lowerCAmelCase: str = 0 try: lowerCAmelCase: Union[str, Any] = proth(number) except ValueError: print(F"ValueError: there is no {number}th Proth number") continue print(F"The {number}th Proth number: {value}")
351
'''simple docstring''' from math import factorial, pi def lowerCamelCase__ ( _A , _A = 30 ): if not isinstance(_A , (int, float) ): raise ValueError('maclaurin_sin() requires either an int or float for theta' ) if not isinstance(_A , _A ) or accuracy <= 0: raise ValueError('maclaurin_sin() requires a positive int for accuracy' ) a : Dict = float(_A ) a : List[Any] = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(_A ) ) def lowerCamelCase__ ( _A , _A = 30 ): if not isinstance(_A , (int, float) ): raise ValueError('maclaurin_cos() requires either an int or float for theta' ) if not isinstance(_A , _A ) or accuracy <= 0: raise ValueError('maclaurin_cos() requires a positive int for accuracy' ) a : int = float(_A ) a : str = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(_A ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(1_0)) print(maclaurin_sin(-1_0)) print(maclaurin_sin(1_0, 1_5)) print(maclaurin_sin(-1_0, 1_5)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(1_0, 1_5)) print(maclaurin_cos(-1_0, 1_5))
96
0
import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() __snake_case : str = logging.get_logger(__name__) __snake_case : int = [ ("""bert.bert""", """visual_bert"""), ("""bert.cls""", """cls"""), ("""bert.classifier""", """cls"""), ("""token_type_embeddings_visual""", """visual_token_type_embeddings"""), ("""position_embeddings_visual""", """visual_position_embeddings"""), ("""projection""", """visual_projection"""), ] __snake_case : List[Any] = [ """nlvr2_coco_pre_trained.th""", """nlvr2_fine_tuned.th""", """nlvr2_pre_trained.th""", """vcr_coco_pre_train.th""", """vcr_fine_tune.th""", """vcr_pre_train.th""", """vqa_coco_pre_trained.th""", """vqa_fine_tuned.th""", """vqa_pre_trained.th""", ] def _UpperCAmelCase ( a__): '''simple docstring''' a_ : Any = torch.load(a__ , map_location="""cpu""") return sd def _UpperCAmelCase ( a__ , a__ , a__=rename_keys_prefix): '''simple docstring''' a_ : str = OrderedDict() a_ : Tuple = torch.arange(config.max_position_embeddings).expand((1, -1)) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue a_ : int = key for name_pair in rename_keys_prefix: a_ : Dict = new_key.replace(name_pair[0] , name_pair[1]) a_ : Dict = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately a_ : Any = new_d["""cls.predictions.bias"""] return new_d @torch.no_grad() def _UpperCAmelCase ( a__ , a__): '''simple docstring''' assert ( checkpoint_path.split("""/""")[-1] in ACCEPTABLE_CHECKPOINTS ), f'''The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.''' # Get Config if "pre" in checkpoint_path: a_ : Any = """pretraining""" if "vcr" in checkpoint_path: a_ : Dict = {"""visual_embedding_dim""": 5_1_2} elif "vqa_advanced" in checkpoint_path: a_ : Tuple = {"""visual_embedding_dim""": 2_0_4_8} elif "vqa" in checkpoint_path: a_ : List[str] = {"""visual_embedding_dim""": 2_0_4_8} elif "nlvr" in checkpoint_path: a_ : int = {"""visual_embedding_dim""": 1_0_2_4} else: raise NotImplementedError(f'''No implementation found for `{checkpoint_path}`.''') else: if "vcr" in checkpoint_path: a_ : List[str] = {"""visual_embedding_dim""": 5_1_2} a_ : Union[str, Any] = """multichoice""" elif "vqa_advanced" in checkpoint_path: a_ : List[Any] = {"""visual_embedding_dim""": 2_0_4_8} a_ : str = """vqa_advanced""" elif "vqa" in checkpoint_path: a_ : str = {"""visual_embedding_dim""": 2_0_4_8, """num_labels""": 3_1_2_9} a_ : Optional[int] = """vqa""" elif "nlvr" in checkpoint_path: a_ : Tuple = { """visual_embedding_dim""": 1_0_2_4, """num_labels""": 2, } a_ : List[str] = """nlvr""" a_ : Dict = VisualBertConfig(**a__) # Load State Dict a_ : List[Any] = load_state_dict(a__) a_ : List[str] = get_new_dict(a__ , a__) if model_type == "pretraining": a_ : Optional[int] = VisualBertForPreTraining(a__) elif model_type == "vqa": a_ : Optional[int] = VisualBertForQuestionAnswering(a__) elif model_type == "nlvr": a_ : int = VisualBertForVisualReasoning(a__) elif model_type == "multichoice": a_ : Optional[int] = VisualBertForMultipleChoice(a__) model.load_state_dict(a__) # Save Checkpoints Path(a__).mkdir(exist_ok=a__) model.save_pretrained(a__) if __name__ == "__main__": __snake_case : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("""orig_checkpoint_path""", type=str, help="""A path to .th on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", type=str, help="""Path to the output PyTorch model.""") __snake_case : Any = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
248
_lowerCamelCase ={ "joule": 1.0, "kilojoule": 1_0_0_0, "megajoule": 1_0_0_0_0_0_0, "gigajoule": 1_0_0_0_0_0_0_0_0_0, "wattsecond": 1.0, "watthour": 3_6_0_0, "kilowatthour": 3_6_0_0_0_0_0, "newtonmeter": 1.0, "calorie_nutr": 4_1_8_6.8, "kilocalorie_nutr": 4_1_8_6_8_0_0.0_0, "electronvolt": 1.6_0_2_1_7_6_6_3_4E-1_9, "britishthermalunit_it": 1_0_5_5.0_5_5_8_5, "footpound": 1.355818, } def _a ( lowerCamelCase, lowerCamelCase, lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: lowerCamelCase : Dict = ( F'''Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n''' F'''Valid values are: {", ".join(lowerCamelCase )}''' ) raise ValueError(lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
287
0
import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": lowerCamelCase : Optional[Any] =pd.read_csv('''sample_data.csv''', header=None) lowerCamelCase : Any =df.shape[:1][0] # If you're using some other dataset input the target column lowerCamelCase : Any =df.iloc[:, 1:2] lowerCamelCase : str =actual_data.values.reshape(len_data, 1) lowerCamelCase : Dict =MinMaxScaler().fit_transform(actual_data) lowerCamelCase : List[str] =10 lowerCamelCase : List[str] =5 lowerCamelCase : int =20 lowerCamelCase : Tuple =len_data - periods * look_back lowerCamelCase : List[Any] =actual_data[:division] lowerCamelCase : List[Any] =actual_data[division - look_back :] lowerCamelCase , lowerCamelCase : List[str] =[], [] lowerCamelCase , lowerCamelCase : int =[], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) lowerCamelCase : List[Any] =np.array(train_x) lowerCamelCase : str =np.array(test_x) lowerCamelCase : Optional[Any] =np.array([list(i.ravel()) for i in train_y]) lowerCamelCase : Any =np.array([list(i.ravel()) for i in test_y]) lowerCamelCase : str =Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss='''mean_squared_error''', optimizer='''adam''') lowerCamelCase : Tuple =model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) lowerCamelCase : List[Any] =model.predict(x_test)
355
from ..utils import DummyObject, requires_backends class __a ( metaclass=A__ ): _lowerCAmelCase : str = ['''torch'''] def __init__( self : int , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[Any] = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Union[str, Any] = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : str , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[str] = ['''torch'''] def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : List[Any] , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : List[Any] , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Any = ['''torch'''] def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : Optional[int] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[Any] = ['''torch'''] def __init__( self : int , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[Any] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Optional[int] = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' requires_backends(cls , ["torch"] ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> int: requires_backends(__lowerCAmelCase , ["torch"] ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Optional[Any]: requires_backends(__lowerCAmelCase , ["torch"] ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Optional[Any]: requires_backends(__lowerCAmelCase , ["torch"] ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Union[str, Any]: requires_backends(__lowerCAmelCase , ["torch"] ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Any: requires_backends(__lowerCAmelCase , ["torch"] ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Dict: requires_backends(__lowerCAmelCase , ["torch"] ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[Any]: requires_backends(__lowerCAmelCase , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[str] = ['''torch'''] def __init__( self : Tuple , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Optional[int] = ['''torch'''] def __init__( self : List[Any] , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Union[str, Any] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : str = ['''torch'''] def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Any = ['''torch'''] def __init__( self : List[Any] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[Any] , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : str = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Any = ['''torch'''] def __init__( self : str , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : int = ['''torch'''] def __init__( self : Optional[Any] , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Union[str, Any] , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Optional[int] = ['''torch'''] def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Union[str, Any] , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[str] = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : Optional[int] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Any = ['''torch'''] def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Any = ['''torch'''] def __init__( self : Optional[int] , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : str , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Union[str, Any] = ['''torch'''] def __init__( self : str , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Any = ['''torch'''] def __init__( self : int , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Union[str, Any] = ['''torch'''] def __init__( self : Optional[Any] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Optional[int] = ['''torch'''] def __init__( self : int , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[Any] = ['''torch'''] def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Optional[int] = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Union[str, Any] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : str = ['''torch'''] def __init__( self : List[str] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : List[Any] , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Optional[int] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : str , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Union[str, Any] , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : Optional[Any] , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : Optional[int] , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : str = ['''torch'''] def __init__( self : Tuple , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[Any] = ['''torch'''] def __init__( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Union[str, Any] , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Union[str, Any] = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : str = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : Optional[Any] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Union[str, Any] = ['''torch'''] def __init__( self : Dict , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : List[str] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Any , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Union[str, Any] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : Tuple ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : str , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[Any] = ['''torch'''] def __init__( self : Tuple , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : Dict , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Optional[int] = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : int , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : List[str] = ['''torch'''] def __init__( self : int , *SCREAMING_SNAKE_CASE : Tuple , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : List[str] , *SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : List[Any] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : Dict ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Tuple = ['''torch'''] def __init__( self : Any , *SCREAMING_SNAKE_CASE : str , **SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Tuple , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Optional[int] , *SCREAMING_SNAKE_CASE : Optional[Any] , **SCREAMING_SNAKE_CASE : int ): '''simple docstring''' requires_backends(cls , ["torch"] ) class __a ( metaclass=A__ ): _lowerCAmelCase : Dict = ['''torch'''] def __init__( self : Optional[int] , *SCREAMING_SNAKE_CASE : List[Any] , **SCREAMING_SNAKE_CASE : str ): '''simple docstring''' requires_backends(self , ["torch"] ) @classmethod def __lowercase ( cls : Optional[Any] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Union[str, Any] ): '''simple docstring''' requires_backends(cls , ["torch"] ) @classmethod def __lowercase ( cls : Dict , *SCREAMING_SNAKE_CASE : Union[str, Any] , **SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' requires_backends(cls , ["torch"] )
196
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class _SCREAMING_SNAKE_CASE( A , A , A , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Tuple = StableDiffusionInstructPixaPixPipeline SCREAMING_SNAKE_CASE_ : Tuple = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} SCREAMING_SNAKE_CASE_ : Any = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS SCREAMING_SNAKE_CASE_ : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS SCREAMING_SNAKE_CASE_ : str = IMAGE_TO_IMAGE_IMAGE_PARAMS def _UpperCamelCase ( self ) -> int: """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE :Any = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=8 ,out_channels=4 ,down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') ,up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') ,cross_attention_dim=32 ,) __SCREAMING_SNAKE_CASE :Optional[int] = PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE :List[Any] = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] ,up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] ,latent_channels=4 ,) torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE :Optional[int] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) __SCREAMING_SNAKE_CASE :Tuple = CLIPTextModel(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Any = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) __SCREAMING_SNAKE_CASE :Tuple = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__=0 ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = floats_tensor((1, 3, 32, 32) ,rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Dict = image.cpu().permute(0 ,2 ,3 ,1 )[0] __SCREAMING_SNAKE_CASE :Tuple = Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('''RGB''' ) if str(lowerCamelCase__ ).startswith('''mps''' ): __SCREAMING_SNAKE_CASE :Dict = torch.manual_seed(lowerCamelCase__ ) else: __SCREAMING_SNAKE_CASE :str = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :List[Any] = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :Tuple = self.get_dummy_components() __SCREAMING_SNAKE_CASE :List[str] = StableDiffusionInstructPixaPixPipeline(**lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :List[Any] = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Optional[int] = self.get_dummy_inputs(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :List[str] = sd_pipe(**lowerCamelCase__ ).images __SCREAMING_SNAKE_CASE :List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE :Dict = np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE :Union[str, Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :int = self.get_dummy_components() __SCREAMING_SNAKE_CASE :str = StableDiffusionInstructPixaPixPipeline(**lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :str = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Optional[int] = self.get_dummy_inputs(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = 'french fries' __SCREAMING_SNAKE_CASE :Union[str, Any] = sd_pipe(**lowerCamelCase__ ,negative_prompt=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :int = output.images __SCREAMING_SNAKE_CASE :Any = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE :Tuple = np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = 'cpu' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :List[Any] = self.get_dummy_components() __SCREAMING_SNAKE_CASE :List[str] = StableDiffusionInstructPixaPixPipeline(**lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Optional[int] = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :int = self.get_dummy_inputs(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :List[str] = [inputs['prompt']] * 2 __SCREAMING_SNAKE_CASE :List[str] = np.array(inputs['''image'''] ).astype(np.floataa ) / 2_5_5.0 __SCREAMING_SNAKE_CASE :Any = torch.from_numpy(lowerCamelCase__ ).unsqueeze(0 ).to(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :int = image / 2 + 0.5 __SCREAMING_SNAKE_CASE :Dict = image.permute(0 ,3 ,1 ,2 ) __SCREAMING_SNAKE_CASE :Dict = image.repeat(2 ,1 ,1 ,1 ) __SCREAMING_SNAKE_CASE :str = sd_pipe(**lowerCamelCase__ ).images __SCREAMING_SNAKE_CASE :Optional[Any] = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) __SCREAMING_SNAKE_CASE :List[Any] = np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> Any: """simple docstring""" __SCREAMING_SNAKE_CASE :Any = 'cpu' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :Any = self.get_dummy_components() __SCREAMING_SNAKE_CASE :List[Any] = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 ,beta_end=0.0_1_2 ,beta_schedule='''scaled_linear''' ) __SCREAMING_SNAKE_CASE :Tuple = StableDiffusionInstructPixaPixPipeline(**lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Dict = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :List[str] = self.get_dummy_inputs(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Optional[Any] = sd_pipe(**lowerCamelCase__ ).images __SCREAMING_SNAKE_CASE :Optional[int] = image[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :Union[str, Any] = [round(lowerCamelCase__ ,4 ) for x in image_slice.flatten().tolist()] print(''','''.join([str(lowerCamelCase__ ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) __SCREAMING_SNAKE_CASE :Optional[Any] = np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> Dict: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = self.get_dummy_components() __SCREAMING_SNAKE_CASE :List[str] = StableDiffusionInstructPixaPixPipeline(**lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = VaeImageProcessor(do_resize=lowerCamelCase__ ,do_normalize=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Optional[int] = pipe(**self.get_dummy_inputs_by_type(lowerCamelCase__ ,input_image_type='''pt''' ) )[0] __SCREAMING_SNAKE_CASE :Union[str, Any] = components['vae'] __SCREAMING_SNAKE_CASE :int = self.get_dummy_inputs_by_type(lowerCamelCase__ ,input_image_type='''pt''' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __SCREAMING_SNAKE_CASE :List[str] = vae.encode(inputs[image_param] ).latent_dist.mode() __SCREAMING_SNAKE_CASE :Union[str, Any] = pipe(**lowerCamelCase__ )[0] __SCREAMING_SNAKE_CASE :Optional[int] = np.abs(out - out_latents_inputs ).max() self.assertLess(lowerCamelCase__ ,1E-4 ,'''passing latents as image input generate different result from passing image''' ) @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE( unittest.TestCase ): def _UpperCamelCase ( self ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__=0 ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = torch.manual_seed(lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Tuple = load_image( '''https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg''' ) __SCREAMING_SNAKE_CASE :str = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' ,safety_checker=lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE :int = self.get_inputs() __SCREAMING_SNAKE_CASE :Optional[Any] = pipe(**lowerCamelCase__ ).images __SCREAMING_SNAKE_CASE :Optional[int] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __SCREAMING_SNAKE_CASE :str = np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' ,safety_checker=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Tuple = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE :List[Any] = self.get_inputs() __SCREAMING_SNAKE_CASE :Optional[int] = pipe(**lowerCamelCase__ ).images __SCREAMING_SNAKE_CASE :Dict = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __SCREAMING_SNAKE_CASE :Union[str, Any] = np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> Tuple: """simple docstring""" __SCREAMING_SNAKE_CASE :int = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' ,safety_checker=lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :List[Any] = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE :Tuple = self.get_inputs() __SCREAMING_SNAKE_CASE :List[str] = pipe(**lowerCamelCase__ ).images __SCREAMING_SNAKE_CASE :int = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __SCREAMING_SNAKE_CASE :str = np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :str = 0 def callback_fn(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> None: __SCREAMING_SNAKE_CASE :List[Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __SCREAMING_SNAKE_CASE :Union[str, Any] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __SCREAMING_SNAKE_CASE :List[Any] = latents[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :Tuple = np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 elif step == 2: __SCREAMING_SNAKE_CASE :str = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __SCREAMING_SNAKE_CASE :Dict = latents[0, -3:, -3:, -1] __SCREAMING_SNAKE_CASE :Any = np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 __SCREAMING_SNAKE_CASE :Dict = False __SCREAMING_SNAKE_CASE :Any = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' ,safety_checker=lowerCamelCase__ ,torch_dtype=torch.floataa ) __SCREAMING_SNAKE_CASE :Tuple = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE :Any = self.get_inputs() pipe(**lowerCamelCase__ ,callback=lowerCamelCase__ ,callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def _UpperCamelCase ( self ) -> List[str]: """simple docstring""" torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __SCREAMING_SNAKE_CASE :List[Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' ,safety_checker=lowerCamelCase__ ,torch_dtype=torch.floataa ) __SCREAMING_SNAKE_CASE :Dict = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __SCREAMING_SNAKE_CASE :str = self.get_inputs() __SCREAMING_SNAKE_CASE :Optional[int] = pipe(**lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Union[str, Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def _UpperCamelCase ( self ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE :List[Any] = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __SCREAMING_SNAKE_CASE :List[Any] = inputs['image'].resize((5_04, 5_04) ) __SCREAMING_SNAKE_CASE :Tuple = 'timbrooks/instruct-pix2pix' __SCREAMING_SNAKE_CASE :List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( lowerCamelCase__ ,safety_checker=lowerCamelCase__ ,) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() __SCREAMING_SNAKE_CASE :Optional[int] = pipe(**lowerCamelCase__ ) __SCREAMING_SNAKE_CASE :Tuple = output.images[0] __SCREAMING_SNAKE_CASE :Optional[Any] = image[2_55:2_58, 3_83:3_86, -1] assert image.shape == (5_04, 5_04, 3) __SCREAMING_SNAKE_CASE :Dict = np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3
191
import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMInverseScheduler, DDIMScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, StableDiffusionDiffEditPipeline, UNetaDConditionModel, ) from diffusers.utils import load_image, slow from diffusers.utils.testing_utils import enable_full_determinism, floats_tensor, require_torch_gpu, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __A ( a , a , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Optional[Any] =StableDiffusionDiffEditPipeline UpperCamelCase__ : str =TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {"""height""", """width""", """image"""} | {"""image_latents"""} UpperCamelCase__ : Optional[Any] =TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS - {"""image"""} | {"""image_latents"""} UpperCamelCase__ : Dict =frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess UpperCamelCase__ : Any =frozenset([] ) def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Dict =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase__ , ) __UpperCamelCase : List[str] =DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) __UpperCamelCase : Union[str, Any] =DDIMInverseScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_zero=lowerCamelCase__ , ) torch.manual_seed(0 ) __UpperCamelCase : Optional[int] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __UpperCamelCase : Tuple =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) __UpperCamelCase : Any =CLIPTextModel(lowerCamelCase__ ) __UpperCamelCase : int =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __UpperCamelCase : Union[str, Any] ={ 'unet': unet, 'scheduler': scheduler, 'inverse_scheduler': inverse_scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : int =floats_tensor((1, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =floats_tensor((1, 2, 4, 16, 16) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Optional[int] =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Dict ={ 'prompt': 'a dog and a newt', 'mask_image': mask, 'image_latents': latents, 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : Tuple =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : int =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : Optional[Any] =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : List[Any] =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : Any =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'source_prompt': 'a cat and a frog', 'target_prompt': 'a dog and a newt', 'generator': generator, 'num_inference_steps': 2, 'num_maps_per_mask': 2, 'mask_encode_strength': 1.0, 'guidance_scale': 6.0, 'output_type': 'numpy', } return inputs def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=0 ): """simple docstring""" __UpperCamelCase : str =floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __UpperCamelCase : Any =image.cpu().permute(0 , 2 , 3 , 1 )[0] __UpperCamelCase : int =Image.fromarray(np.uinta(lowerCamelCase__ ) ).convert('RGB' ) if str(lowerCamelCase__ ).startswith('mps' ): __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) else: __UpperCamelCase : int =torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[int] ={ 'image': image, 'prompt': 'a cat and a frog', 'generator': generator, 'num_inference_steps': 2, 'inpaint_strength': 1.0, 'guidance_scale': 6.0, 'decode_latents': True, 'output_type': 'numpy', } return inputs def __lowercase ( self ): """simple docstring""" if not hasattr(self.pipeline_class , '_optional_components' ): return __UpperCamelCase : Optional[Any] =self.get_dummy_components() __UpperCamelCase : List[str] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) # set all optional components to None and update pipeline config accordingly for optional_component in pipe._optional_components: setattr(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) pipe.register_modules(**{optional_component: None for optional_component in pipe._optional_components} ) __UpperCamelCase : Union[str, Any] =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe(**lowerCamelCase__ )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : Tuple =self.pipeline_class.from_pretrained(lowerCamelCase__ ) pipe_loaded.to(lowerCamelCase__ ) pipe_loaded.set_progress_bar_config(disable=lowerCamelCase__ ) for optional_component in pipe._optional_components: self.assertTrue( getattr(lowerCamelCase__ , lowerCamelCase__ ) is None , f'`{optional_component}` did not stay set to None after loading.' , ) __UpperCamelCase : str =self.get_dummy_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe_loaded(**lowerCamelCase__ )[0] __UpperCamelCase : Tuple =np.abs(output - output_loaded ).max() self.assertLess(lowerCamelCase__ , 1E-4 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : int =self.get_dummy_mask_inputs(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =pipe.generate_mask(**lowerCamelCase__ ) __UpperCamelCase : int =mask[0, -3:, -3:] self.assertEqual(mask.shape , (1, 16, 16) ) __UpperCamelCase : Tuple =np.array([0] * 9 ) __UpperCamelCase : str =np.abs(mask_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) self.assertEqual(mask[0, -3, -4] , 0 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int ='cpu' __UpperCamelCase : Union[str, Any] =self.get_dummy_components() __UpperCamelCase : Optional[Any] =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Dict =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : List[Any] =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : Optional[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : int =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) def __lowercase ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=5E-3 ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ='cpu' __UpperCamelCase : int =self.get_dummy_components() __UpperCamelCase : str ={'beta_start': 0.00_085, 'beta_end': 0.012, 'beta_schedule': 'scaled_linear'} __UpperCamelCase : str =DPMSolverMultistepScheduler(**lowerCamelCase__ ) __UpperCamelCase : Dict =DPMSolverMultistepInverseScheduler(**lowerCamelCase__ ) __UpperCamelCase : Any =self.pipeline_class(**lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Tuple =self.get_dummy_inversion_inputs(lowerCamelCase__ ) __UpperCamelCase : str =pipe.invert(**lowerCamelCase__ ).images __UpperCamelCase : List[Any] =image[0, -1, -3:, -3:] self.assertEqual(image.shape , (2, 32, 32, 3) ) __UpperCamelCase : List[str] =np.array( [0.5_150, 0.5_134, 0.5_043, 0.5_376, 0.4_694, 0.51_050, 0.5_015, 0.4_407, 0.4_799] , ) __UpperCamelCase : Optional[Any] =np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase__ , 1E-3 ) @require_torch_gpu @slow class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @classmethod def __lowercase ( cls ): """simple docstring""" __UpperCamelCase : Optional[int] =load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/diffedit/fruit.png' ) __UpperCamelCase : Union[str, Any] =raw_image.convert('RGB' ).resize((768, 768) ) __UpperCamelCase : List[Any] =raw_image def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =torch.manual_seed(0 ) __UpperCamelCase : Dict =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : List[str] =DDIMScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : List[str] =DDIMInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : List[str] ='a bowl of fruit' __UpperCamelCase : Dict ='a bowl of pears' __UpperCamelCase : Tuple =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : int =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ ).latents __UpperCamelCase : Dict =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , output_type='numpy' , ).images[0] __UpperCamelCase : str =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =torch.manual_seed(0 ) __UpperCamelCase : List[Any] =StableDiffusionDiffEditPipeline.from_pretrained( 'stabilityai/stable-diffusion-2-1' , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa ) __UpperCamelCase : Optional[Any] =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) __UpperCamelCase : Optional[int] =DPMSolverMultistepInverseScheduler.from_config(pipe.scheduler.config ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] ='a bowl of fruit' __UpperCamelCase : int ='a bowl of pears' __UpperCamelCase : str =pipe.generate_mask( image=self.raw_image , source_prompt=lowerCamelCase__ , target_prompt=lowerCamelCase__ , generator=lowerCamelCase__ , ) __UpperCamelCase : List[str] =pipe.invert( prompt=lowerCamelCase__ , image=self.raw_image , inpaint_strength=0.7 , generator=lowerCamelCase__ , num_inference_steps=25 , ).latents __UpperCamelCase : List[str] =pipe( prompt=lowerCamelCase__ , mask_image=lowerCamelCase__ , image_latents=lowerCamelCase__ , generator=lowerCamelCase__ , negative_prompt=lowerCamelCase__ , inpaint_strength=0.7 , num_inference_steps=25 , output_type='numpy' , ).images[0] __UpperCamelCase : Tuple =( np.array( load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/diffedit/pears.png' ).resize((768, 768) ) ) / 255 ) assert np.abs((expected_image - image).max() ) < 5E-1
71
0
import random from typing import Any def _snake_case( SCREAMING_SNAKE_CASE__ : list ) -> list[Any]: '''simple docstring''' for _ in range(len(SCREAMING_SNAKE_CASE__ ) ): A__ = random.randint(0 , len(SCREAMING_SNAKE_CASE__ ) - 1 ) A__ = random.randint(0 , len(SCREAMING_SNAKE_CASE__ ) - 1 ) A__ , A__ = data[b], data[a] return data if __name__ == "__main__": lowercase_ = [0, 1, 2, 3, 4, 5, 6, 7] lowercase_ = ["python", "says", "hello", "!"] print("Fisher-Yates Shuffle:") print("List", integers, strings) print("FY Shuffle", fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
282
from __future__ import annotations from collections.abc import Iterable, Iterator from dataclasses import dataclass lowercase_ = (3, 9, -11, 0, 7, 5, 1, -1) lowercase_ = (4, 6, 2, 0, 8, 10, 3, -2) @dataclass class A : """simple docstring""" lowerCamelCase = 42 lowerCamelCase = 42 class A : """simple docstring""" def __init__( self : List[str],lowercase_ : Iterable[int] )-> None: '''simple docstring''' A__ = None for i in sorted(lowercase_,reverse=lowercase_ ): A__ = Node(lowercase_,self.head ) def __iter__( self : List[str] )-> Iterator[int]: '''simple docstring''' A__ = self.head while node: yield node.data A__ = node.next_node def __len__( self : str )-> int: '''simple docstring''' return sum(1 for _ in self ) def __str__( self : Optional[int] )-> str: '''simple docstring''' return " -> ".join([str(lowercase_ ) for node in self] ) def _snake_case( SCREAMING_SNAKE_CASE__ : SortedLinkedList , SCREAMING_SNAKE_CASE__ : SortedLinkedList ) -> SortedLinkedList: '''simple docstring''' return SortedLinkedList(list(SCREAMING_SNAKE_CASE__ ) + list(SCREAMING_SNAKE_CASE__ ) ) if __name__ == "__main__": import doctest doctest.testmod() lowercase_ = SortedLinkedList print(merge_lists(SSL(test_data_odd), SSL(test_data_even)))
282
1
"""simple docstring""" def lowercase (SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : int ) -> float: if digit_amount > 0: return round(number - int(_lowerCAmelCase ) , _lowerCAmelCase ) return number - int(_lowerCAmelCase ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
113
import torch import torch.nn as nn from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel from ...utils import logging _lowerCAmelCase : List[str] = logging.get_logger(__name__) def __snake_case ( _lowerCAmelCase : int , _lowerCAmelCase : Any ) -> Dict: A_ : Optional[Any] = nn.functional.normalize(_lowerCAmelCase ) A_ : List[str] = nn.functional.normalize(_lowerCAmelCase ) return torch.mm(_lowerCAmelCase , normalized_text_embeds.t() ) class __magic_name__ ( lowerCamelCase__ ): """simple docstring""" __UpperCamelCase = CLIPConfig __UpperCamelCase = ['''CLIPEncoderLayer'''] def __init__( self :int , snake_case :CLIPConfig ): '''simple docstring''' super().__init__(snake_case ) A_ : int = CLIPVisionModel(config.vision_config ) A_ : List[str] = nn.Linear(config.vision_config.hidden_size , config.projection_dim , bias=snake_case ) A_ : Tuple = nn.Parameter(torch.ones(17 , config.projection_dim ) , requires_grad=snake_case ) A_ : str = nn.Parameter(torch.ones(3 , config.projection_dim ) , requires_grad=snake_case ) A_ : List[str] = nn.Parameter(torch.ones(17 ) , requires_grad=snake_case ) A_ : int = nn.Parameter(torch.ones(3 ) , requires_grad=snake_case ) @torch.no_grad() def SCREAMING_SNAKE_CASE ( self :Union[str, Any] , snake_case :Dict , snake_case :Any ): '''simple docstring''' A_ : List[Any] = self.vision_model(snake_case )[1] # pooled_output A_ : List[Any] = self.visual_projection(snake_case ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 A_ : Optional[Any] = cosine_distance(snake_case , self.special_care_embeds ).cpu().float().numpy() A_ : Tuple = cosine_distance(snake_case , self.concept_embeds ).cpu().float().numpy() A_ : Union[str, Any] = [] A_ : Any = image_embeds.shape[0] for i in range(snake_case ): A_ : Optional[int] = {"special_scores": {}, "special_care": [], "concept_scores": {}, "bad_concepts": []} # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign images A_ : Optional[Any] = 0.0 for concept_idx in range(len(special_cos_dist[0] ) ): A_ : Optional[Any] = special_cos_dist[i][concept_idx] A_ : Tuple = self.special_care_embeds_weights[concept_idx].item() A_ : Union[str, Any] = round(concept_cos - concept_threshold + adjustment , 3 ) if result_img["special_scores"][concept_idx] > 0: result_img["special_care"].append({concept_idx, result_img["special_scores"][concept_idx]} ) A_ : Any = 0.01 for concept_idx in range(len(cos_dist[0] ) ): A_ : Tuple = cos_dist[i][concept_idx] A_ : Tuple = self.concept_embeds_weights[concept_idx].item() A_ : Tuple = round(concept_cos - concept_threshold + adjustment , 3 ) if result_img["concept_scores"][concept_idx] > 0: result_img["bad_concepts"].append(snake_case ) result.append(snake_case ) A_ : Any = [len(res["bad_concepts"] ) > 0 for res in result] return images, has_nsfw_concepts @torch.no_grad() def SCREAMING_SNAKE_CASE ( self :Union[str, Any] , snake_case :torch.FloatTensor , snake_case :torch.FloatTensor ): '''simple docstring''' A_ : List[str] = self.vision_model(snake_case )[1] # pooled_output A_ : int = self.visual_projection(snake_case ) A_ : Tuple = cosine_distance(snake_case , self.special_care_embeds ) A_ : Tuple = cosine_distance(snake_case , self.concept_embeds ) # increase this value to create a stronger `nsfw` filter # at the cost of increasing the possibility of filtering benign images A_ : Optional[Any] = 0.0 A_ : Tuple = special_cos_dist - self.special_care_embeds_weights + adjustment # special_scores = special_scores.round(decimals=3) A_ : Optional[Any] = torch.any(special_scores > 0 , dim=1 ) A_ : Optional[Any] = special_care * 0.01 A_ : Optional[int] = special_adjustment.unsqueeze(1 ).expand(-1 , cos_dist.shape[1] ) A_ : Union[str, Any] = (cos_dist - self.concept_embeds_weights) + special_adjustment # concept_scores = concept_scores.round(decimals=3) A_ : Union[str, Any] = torch.any(concept_scores > 0 , dim=1 ) return images, has_nsfw_concepts
300
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) class lowercase ( __lowerCamelCase ): __SCREAMING_SNAKE_CASE : int = ['pixel_values'] def __init__( self , snake_case = True , snake_case = None , snake_case = PILImageResampling.BICUBIC , snake_case = True , snake_case = 1 / 255 , snake_case = True , snake_case = None , snake_case = None , snake_case = True , **snake_case , ): super().__init__(**UpperCamelCase_ ) snake_case_ = size if size is not None else {'height': 384, 'width': 384} snake_case_ = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) snake_case_ = do_resize snake_case_ = size snake_case_ = resample snake_case_ = do_rescale snake_case_ = rescale_factor snake_case_ = do_normalize snake_case_ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN snake_case_ = image_std if image_std is not None else OPENAI_CLIP_STD snake_case_ = do_convert_rgb def a ( self , snake_case , snake_case , snake_case = PILImageResampling.BICUBIC , snake_case = None , **snake_case , ): snake_case_ = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) if "height" not in size or "width" not in size: raise ValueError(F'''The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}''' ) snake_case_ = (size['height'], size['width']) return resize(UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def a ( self , snake_case , snake_case , snake_case = None , **snake_case , ): return rescale(UpperCamelCase_ , scale=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def a ( self , snake_case , snake_case , snake_case , snake_case = None , **snake_case , ): return normalize(UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ , data_format=UpperCamelCase_ , **UpperCamelCase_ ) def a ( self , snake_case , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = ChannelDimension.FIRST , **snake_case , ): snake_case_ = do_resize if do_resize is not None else self.do_resize snake_case_ = resample if resample is not None else self.resample snake_case_ = do_rescale if do_rescale is not None else self.do_rescale snake_case_ = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ = do_normalize if do_normalize is not None else self.do_normalize snake_case_ = image_mean if image_mean is not None else self.image_mean snake_case_ = image_std if image_std is not None else self.image_std snake_case_ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb snake_case_ = size if size is not None else self.size snake_case_ = get_size_dict(UpperCamelCase_ , default_to_square=UpperCamelCase_ ) snake_case_ = make_list_of_images(UpperCamelCase_ ) if not valid_images(UpperCamelCase_ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: snake_case_ = [convert_to_rgb(UpperCamelCase_ ) for image in images] # All transformations expect numpy arrays. snake_case_ = [to_numpy_array(UpperCamelCase_ ) for image in images] if do_resize: snake_case_ = [self.resize(image=UpperCamelCase_ , size=UpperCamelCase_ , resample=UpperCamelCase_ ) for image in images] if do_rescale: snake_case_ = [self.rescale(image=UpperCamelCase_ , scale=UpperCamelCase_ ) for image in images] if do_normalize: snake_case_ = [self.normalize(image=UpperCamelCase_ , mean=UpperCamelCase_ , std=UpperCamelCase_ ) for image in images] snake_case_ = [to_channel_dimension_format(UpperCamelCase_ , UpperCamelCase_ ) for image in images] snake_case_ = BatchFeature(data={'pixel_values': images} , tensor_type=UpperCamelCase_ ) return encoded_outputs
357
import gc import unittest from diffusers import FlaxControlNetModel, FlaxStableDiffusionControlNetPipeline from diffusers.utils import is_flax_available, load_image, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class lowercase ( unittest.TestCase ): def a ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() def a ( self ): snake_case_ , snake_case_ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-canny' , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ , snake_case_ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=snake_case , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ = controlnet_params snake_case_ = 'bird' snake_case_ = jax.device_count() snake_case_ = pipe.prepare_text_inputs([prompts] * num_samples ) snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png' ) snake_case_ = pipe.prepare_image_inputs([canny_image] * num_samples ) snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = jax.random.split(snake_case , jax.device_count() ) snake_case_ = replicate(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = pipe( prompt_ids=snake_case , image=snake_case , params=snake_case , prng_seed=snake_case , num_inference_steps=50 , jit=snake_case , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) snake_case_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) snake_case_ = images[0, 253:256, 253:256, -1] snake_case_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) snake_case_ = jnp.array( [0.16_79_69, 0.11_66_99, 0.08_15_43, 0.15_42_97, 0.13_28_12, 0.10_88_87, 0.16_99_22, 0.16_99_22, 0.20_50_78] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def a ( self ): snake_case_ , snake_case_ = FlaxControlNetModel.from_pretrained( 'lllyasviel/sd-controlnet-openpose' , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ , snake_case_ = FlaxStableDiffusionControlNetPipeline.from_pretrained( 'runwayml/stable-diffusion-v1-5' , controlnet=snake_case , from_pt=snake_case , dtype=jnp.bfloataa ) snake_case_ = controlnet_params snake_case_ = 'Chef in the kitchen' snake_case_ = jax.device_count() snake_case_ = pipe.prepare_text_inputs([prompts] * num_samples ) snake_case_ = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/pose.png' ) snake_case_ = pipe.prepare_image_inputs([pose_image] * num_samples ) snake_case_ = jax.random.PRNGKey(0 ) snake_case_ = jax.random.split(snake_case , jax.device_count() ) snake_case_ = replicate(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = shard(snake_case ) snake_case_ = pipe( prompt_ids=snake_case , image=snake_case , params=snake_case , prng_seed=snake_case , num_inference_steps=50 , jit=snake_case , ).images assert images.shape == (jax.device_count(), 1, 768, 512, 3) snake_case_ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) snake_case_ = images[0, 253:256, 253:256, -1] snake_case_ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) snake_case_ = jnp.array( [[0.27_14_84, 0.26_17_19, 0.27_53_91, 0.27_73_44, 0.27_92_97, 0.29_10_16, 0.29_49_22, 0.30_27_34, 0.30_27_34]] ) print(F'''output_slice: {output_slice}''' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
200
0
'''simple docstring''' from __future__ import annotations __a = 1.6021E-19 # units = C def __snake_case( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) -> tuple[str, float]: if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError("""You cannot supply more or less than 2 values""" ) elif conductivity < 0: raise ValueError("""Conductivity cannot be negative""" ) elif electron_conc < 0: raise ValueError("""Electron concentration cannot be negative""" ) elif mobility < 0: raise ValueError("""mobility cannot be negative""" ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
35
'''simple docstring''' import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class A__ ( unittest.TestCase ): """simple docstring""" UpperCamelCase_ : Any = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def _lowerCAmelCase ( self : Any , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : int ) -> List[Any]: """simple docstring""" _UpperCAmelCase : int = hf_hub_download( repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) _UpperCAmelCase : int = VideoClassificationPipeline(model=lowerCAmelCase__ , image_processor=lowerCAmelCase__ , top_k=2 ) _UpperCAmelCase : int = [ example_video_filepath, "https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4", ] return video_classifier, examples def _lowerCAmelCase ( self : Optional[Any] , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Any ) -> int: """simple docstring""" for example in examples: _UpperCAmelCase : List[str] = video_classifier(lowerCAmelCase__ ) self.assertEqual( lowerCAmelCase__ , [ {"score": ANY(lowerCAmelCase__ ), "label": ANY(lowerCAmelCase__ )}, {"score": ANY(lowerCAmelCase__ ), "label": ANY(lowerCAmelCase__ )}, ] , ) @require_torch def _lowerCAmelCase ( self : Tuple ) -> Union[str, Any]: """simple docstring""" _UpperCAmelCase : List[Any] = "hf-internal-testing/tiny-random-VideoMAEForVideoClassification" _UpperCAmelCase : Optional[int] = VideoMAEFeatureExtractor( size={"shortest_edge": 1_0} , crop_size={"height": 1_0, "width": 1_0} ) _UpperCAmelCase : List[str] = pipeline( "video-classification" , model=lowerCAmelCase__ , feature_extractor=lowerCAmelCase__ , frame_sampling_rate=4 ) _UpperCAmelCase : Tuple = hf_hub_download(repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) _UpperCAmelCase : Tuple = video_classifier(lowerCAmelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}] , ) _UpperCAmelCase : Any = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}], [{"score": 0.5199, "label": "LABEL_0"}, {"score": 0.4801, "label": "LABEL_1"}], ] , ) @require_tf def _lowerCAmelCase ( self : List[str] ) -> int: """simple docstring""" pass
145
0
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. snake_case_ : Any = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. snake_case_ : Any = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. snake_case_ : Optional[int] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Tuple = len([g for position, g in enumerate(UpperCAmelCase_ ) if g == main_target[position]] ) return (item, float(UpperCAmelCase_ )) def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Optional[int] = random.randint(0 , len(UpperCAmelCase_ ) - 1 ) _UpperCamelCase : str = parent_a[:random_slice] + parent_a[random_slice:] _UpperCamelCase : Any = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def A__ ( UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : List[Any] = list(UpperCAmelCase_ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _UpperCamelCase : Union[str, Any] = random.choice(UpperCAmelCase_ ) return "".join(UpperCAmelCase_ ) def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ): _UpperCamelCase : str = [] # Generate more children proportionally to the fitness score. _UpperCamelCase : Dict = int(parent_a[1] * 1_0_0 ) + 1 _UpperCamelCase : Any = 1_0 if child_n >= 1_0 else child_n for _ in range(UpperCAmelCase_ ): _UpperCamelCase : List[Any] = population_score[random.randint(0 , UpperCAmelCase_ )][0] _UpperCamelCase , _UpperCamelCase : Optional[Any] = crossover(parent_a[0] , UpperCAmelCase_ ) # Append new string to the population list. pop.append(mutate(UpperCAmelCase_ , UpperCAmelCase_ ) ) pop.append(mutate(UpperCAmelCase_ , UpperCAmelCase_ ) ) return pop def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _UpperCamelCase : List[str] = f'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(UpperCAmelCase_ ) # Verify that the target contains no genes besides the ones inside genes variable. _UpperCamelCase : Any = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _UpperCamelCase : Dict = f'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(UpperCAmelCase_ ) # Generate random starting population. _UpperCamelCase : int = [] for _ in range(UpperCAmelCase_ ): population.append(''.join([random.choice(UpperCAmelCase_ ) for i in range(len(UpperCAmelCase_ ) )] ) ) # Just some logs to know what the algorithms is doing. _UpperCamelCase , _UpperCamelCase : Any = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(UpperCAmelCase_ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _UpperCamelCase : Tuple = [evaluate(UpperCAmelCase_ , UpperCAmelCase_ ) for item in population] # Check if there is a matching evolution. _UpperCamelCase : Optional[int] = sorted(UpperCAmelCase_ , key=lambda UpperCAmelCase_ : x[1] , reverse=UpperCAmelCase_ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 1_0 == 0: print( f'\nGeneration: {generation}' f'\nTotal Population:{total_population}' f'\nBest score: {population_score[0][1]}' f'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _UpperCamelCase : Dict = population[: int(N_POPULATION / 3 )] population.clear() population.extend(UpperCAmelCase_ ) # Normalize population score to be between 0 and 1. _UpperCamelCase : Any = [ (item, score / len(UpperCAmelCase_ )) for item, score in population_score ] # This is selection for i in range(UpperCAmelCase_ ): population.extend(select(population_score[int(UpperCAmelCase_ )] , UpperCAmelCase_ , UpperCAmelCase_ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(UpperCAmelCase_ ) > N_POPULATION: break if __name__ == "__main__": snake_case_ : List[Any] = ( 'This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!' ) snake_case_ : Tuple = list( ' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm' 'nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\' ) snake_case_ , snake_case_ , snake_case_ : List[str] = basic(target_str, genes_list) print( F"""\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}""" )
236
'''simple docstring''' from __future__ import annotations import numpy as np def A__ ( UpperCAmelCase_ ): _UpperCamelCase , _UpperCamelCase : Optional[int] = np.shape(UpperCAmelCase_ ) if rows != columns: _UpperCamelCase : Union[str, Any] = ( '\'table\' has to be of square shaped array but got a ' f'{rows}x{columns} array:\n{table}' ) raise ValueError(UpperCAmelCase_ ) _UpperCamelCase : Optional[Any] = np.zeros((rows, columns) ) _UpperCamelCase : Tuple = np.zeros((rows, columns) ) for i in range(UpperCAmelCase_ ): for j in range(UpperCAmelCase_ ): _UpperCamelCase : Optional[Any] = sum(lower[i][k] * upper[k][j] for k in range(UpperCAmelCase_ ) ) if upper[j][j] == 0: raise ArithmeticError('No LU decomposition exists' ) _UpperCamelCase : Optional[Any] = (table[i][j] - total) / upper[j][j] _UpperCamelCase : int = 1 for j in range(UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Optional[int] = sum(lower[i][k] * upper[k][j] for k in range(UpperCAmelCase_ ) ) _UpperCamelCase : Tuple = table[i][j] - total return lower, upper if __name__ == "__main__": import doctest doctest.testmod()
236
1
'''simple docstring''' from random import shuffle import tensorflow as tf from numpy import array def _A ( A__ , A__ ): """simple docstring""" __lowercase = int(A__ ) assert noofclusters < len(A__ ) # Find out the dimensionality __lowercase = len(vectors[0] ) # Will help select random centroids from among the available vectors __lowercase = list(range(len(A__ ) ) ) shuffle(A__ ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. __lowercase = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION __lowercase = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points __lowercase = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(A__ ) ] ##These nodes will assign the centroid Variables the appropriate ##values __lowercase = tf.placeholder('''float64''' , [dim] ) __lowercase = [] for centroid in centroids: cent_assigns.append(tf.assign(A__ , A__ ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) __lowercase = [tf.Variable(0 ) for i in range(len(A__ ) )] ##These nodes will assign an assignment Variable the appropriate ##value __lowercase = tf.placeholder('''int32''' ) __lowercase = [] for assignment in assignments: cluster_assigns.append(tf.assign(A__ , A__ ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input __lowercase = tf.placeholder('''float''' , [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors __lowercase = tf.reduce_mean(A__ , 0 ) ##Node for computing Euclidean distances # Placeholders for input __lowercase = tf.placeholder('''float''' , [dim] ) __lowercase = tf.placeholder('''float''' , [dim] ) __lowercase = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(A__ , A__ ) , 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input __lowercase = tf.placeholder('''float''' , [noofclusters] ) __lowercase = tf.argmin(A__ , 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. __lowercase = tf.initialize_all_variables() # Initialize all variables sess.run(A__ ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. __lowercase = 100 for _ in range(A__ ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(A__ ) ): __lowercase = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. __lowercase = [ sess.run(A__ , feed_dict={va: vect, va: sess.run(A__ )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input __lowercase = sess.run( A__ , feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] , feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(A__ ): # Collect all the vectors assigned to this cluster __lowercase = [ vectors[i] for i in range(len(A__ ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location __lowercase = sess.run( A__ , feed_dict={mean_input: array(A__ )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] , feed_dict={centroid_value: new_location} ) # Return centroids and assignments __lowercase = sess.run(A__ ) __lowercase = sess.run(A__ ) return centroids, assignments
104
'''simple docstring''' from __future__ import annotations from random import random from typing import Generic, TypeVar lowerCAmelCase__ = TypeVar('''KT''') lowerCAmelCase__ = TypeVar('''VT''') class lowercase_ (Generic[KT, VT] ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : KT | str = "root" ,lowercase__ : VT | None = None ): __lowercase = key __lowercase = value __lowercase = [] def __repr__( self : Tuple ): return F"Node({self.key}: {self.value})" @property def SCREAMING_SNAKE_CASE ( self : int ): return len(self.forward ) class lowercase_ (Generic[KT, VT] ): """simple docstring""" def __init__( self : int ,lowercase__ : float = 0.5 ,lowercase__ : int = 1_6 ): __lowercase = Node[KT, VT]() __lowercase = 0 __lowercase = p __lowercase = max_level def __str__( self : List[str] ): __lowercase = list(self ) if len(lowercase__ ) == 0: return F"SkipList(level={self.level})" __lowercase = max((len(str(lowercase__ ) ) for item in items) ,default=4 ) __lowercase = max(lowercase__ ,4 ) + 4 __lowercase = self.head __lowercase = [] __lowercase = node.forward.copy() lines.append(F"[{node.key}]".ljust(lowercase__ ,'''-''' ) + '''* ''' * len(lowercase__ ) ) lines.append(''' ''' * label_size + '''| ''' * len(lowercase__ ) ) while len(node.forward ) != 0: __lowercase = node.forward[0] lines.append( F"[{node.key}]".ljust(lowercase__ ,'''-''' ) + ''' '''.join(str(n.key ) if n.key == node.key else '''|''' for n in forwards ) ) lines.append(''' ''' * label_size + '''| ''' * len(lowercase__ ) ) __lowercase = node.forward lines.append('''None'''.ljust(lowercase__ ) + '''* ''' * len(lowercase__ ) ) return F"SkipList(level={self.level})\n" + "\n".join(lowercase__ ) def __iter__( self : List[str] ): __lowercase = self.head while len(node.forward ) != 0: yield node.forward[0].key __lowercase = node.forward[0] def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = 1 while random() < self.p and level < self.max_level: level += 1 return level def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : str ): __lowercase = [] __lowercase = self.head for i in reversed(range(self.level ) ): # i < node.level - When node level is lesser than `i` decrement `i`. # node.forward[i].key < key - Jumping to node with key value higher # or equal to searched key would result # in skipping searched key. while i < node.level and node.forward[i].key < key: __lowercase = node.forward[i] # Each leftmost node (relative to searched node) will potentially have to # be updated. update_vector.append(lowercase__ ) update_vector.reverse() # Note that we were inserting values in reverse order. # len(node.forward) != 0 - If current node doesn't contain any further # references then searched key is not present. # node.forward[0].key == key - Next node key should be equal to search key # if key is present. if len(node.forward ) != 0 and node.forward[0].key == key: return node.forward[0], update_vector else: return None, update_vector def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : KT ): __lowercase , __lowercase = self._locate_node(lowercase__ ) if node is not None: for i, update_node in enumerate(lowercase__ ): # Remove or replace all references to removed node. if update_node.level > i and update_node.forward[i].key == key: if node.level > i: __lowercase = node.forward[i] else: __lowercase = update_node.forward[:i] def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : KT ,lowercase__ : VT ): __lowercase , __lowercase = self._locate_node(lowercase__ ) if node is not None: __lowercase = value else: __lowercase = self.random_level() if level > self.level: # After level increase we have to add additional nodes to head. for _ in range(self.level - 1 ,lowercase__ ): update_vector.append(self.head ) __lowercase = level __lowercase = Node(lowercase__ ,lowercase__ ) for i, update_node in enumerate(update_vector[:level] ): # Change references to pass through new node. if update_node.level > i: new_node.forward.append(update_node.forward[i] ) if update_node.level < i + 1: update_node.forward.append(lowercase__ ) else: __lowercase = new_node def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : VT ): __lowercase , __lowercase = self._locate_node(lowercase__ ) if node is not None: return node.value return None def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.insert('''Key1''' , 3 ) skip_list.insert('''Key2''' , 12 ) skip_list.insert('''Key3''' , 41 ) skip_list.insert('''Key4''' , -19 ) __lowercase = skip_list.head __lowercase = {} while node.level != 0: __lowercase = node.forward[0] __lowercase = node.value assert len(A__ ) == 4 assert all_values["Key1"] == 3 assert all_values["Key2"] == 12 assert all_values["Key3"] == 41 assert all_values["Key4"] == -19 def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.insert('''Key1''' , 10 ) skip_list.insert('''Key1''' , 12 ) skip_list.insert('''Key5''' , 7 ) skip_list.insert('''Key7''' , 10 ) skip_list.insert('''Key10''' , 5 ) skip_list.insert('''Key7''' , 7 ) skip_list.insert('''Key5''' , 5 ) skip_list.insert('''Key10''' , 10 ) __lowercase = skip_list.head __lowercase = {} while node.level != 0: __lowercase = node.forward[0] __lowercase = node.value if len(A__ ) != 4: print() assert len(A__ ) == 4 assert all_values["Key1"] == 12 assert all_values["Key7"] == 7 assert all_values["Key5"] == 5 assert all_values["Key10"] == 10 def _A ( ): """simple docstring""" __lowercase = SkipList() assert skip_list.find('''Some key''' ) is None def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.insert('''Key2''' , 20 ) assert skip_list.find('''Key2''' ) == 20 skip_list.insert('''Some Key''' , 10 ) skip_list.insert('''Key2''' , 8 ) skip_list.insert('''V''' , 13 ) assert skip_list.find('''Y''' ) is None assert skip_list.find('''Key2''' ) == 8 assert skip_list.find('''Some Key''' ) == 10 assert skip_list.find('''V''' ) == 13 def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.delete('''Some key''' ) assert len(skip_list.head.forward ) == 0 def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 14 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''V''' ) skip_list.delete('''Key2''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''Key2''' ) is None def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 14 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''V''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) == 14 assert skip_list.find('''Key1''' ) == 12 assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''X''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) == 12 assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''Key1''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) is None assert skip_list.find('''Key2''' ) == 15 skip_list.delete('''Key2''' ) assert skip_list.find('''V''' ) is None assert skip_list.find('''X''' ) is None assert skip_list.find('''Key1''' ) is None assert skip_list.find('''Key2''' ) is None def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.insert('''Key1''' , 12 ) skip_list.insert('''V''' , 13 ) skip_list.insert('''X''' , 142 ) skip_list.insert('''Key2''' , 15 ) skip_list.delete('''X''' ) def traverse_keys(A__ ): yield node.key for forward_node in node.forward: yield from traverse_keys(A__ ) assert len(set(traverse_keys(skip_list.head ) ) ) == 4 def _A ( ): """simple docstring""" def is_sorted(A__ ): return all(next_item >= item for item, next_item in zip(A__ , lst[1:] ) ) __lowercase = SkipList() for i in range(10 ): skip_list.insert(A__ , A__ ) assert is_sorted(list(A__ ) ) skip_list.delete(5 ) skip_list.delete(8 ) skip_list.delete(2 ) assert is_sorted(list(A__ ) ) skip_list.insert(-12 , -12 ) skip_list.insert(77 , 77 ) assert is_sorted(list(A__ ) ) def _A ( ): """simple docstring""" for _ in range(100 ): # Repeat test 100 times due to the probabilistic nature of skip list # random values == random bugs test_insert() test_insert_overrides_existing_value() test_searching_empty_list_returns_none() test_search() test_deleting_item_from_empty_list_do_nothing() test_deleted_items_are_not_founded_by_find_method() test_delete_removes_only_given_key() test_delete_doesnt_leave_dead_nodes() test_iter_always_yields_sorted_values() def _A ( ): """simple docstring""" __lowercase = SkipList() skip_list.insert(2 , '''2''' ) skip_list.insert(4 , '''4''' ) skip_list.insert(6 , '''4''' ) skip_list.insert(4 , '''5''' ) skip_list.insert(8 , '''4''' ) skip_list.insert(9 , '''4''' ) skip_list.delete(4 ) print(A__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
104
1
'''simple docstring''' import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoFeatureExtractor, WavaVecaFeatureExtractor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / "utils")) from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 __UpperCAmelCase :Any = get_tests_dir("fixtures") class a ( unittest.TestCase ): """simple docstring""" def lowerCamelCase__ ( self : Tuple ) -> List[str]: # A mock response for an HTTP head request to emulate server down __UpperCAmelCase : Dict = mock.Mock() __UpperCAmelCase : List[Any] = 500 __UpperCAmelCase : Dict = {} __UpperCAmelCase : Dict = HTTPError __UpperCAmelCase : List[str] = {} # Download this model to make sure it's in the cache. __UpperCAmelCase : List[str] = WavaVecaFeatureExtractor.from_pretrained('''hf-internal-testing/tiny-random-wav2vec2''' ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('''requests.Session.request''' , return_value=snake_case ) as mock_head: __UpperCAmelCase : Union[str, Any] = WavaVecaFeatureExtractor.from_pretrained('''hf-internal-testing/tiny-random-wav2vec2''' ) # This check we did call the fake head request mock_head.assert_called() def lowerCamelCase__ ( self : List[str] ) -> List[str]: # This test is for deprecated behavior and can be removed in v5 __UpperCAmelCase : List[str] = WavaVecaFeatureExtractor.from_pretrained( '''https://huggingface.co/hf-internal-testing/tiny-random-wav2vec2/resolve/main/preprocessor_config.json''' ) @is_staging_test class a ( unittest.TestCase ): """simple docstring""" @classmethod def lowerCamelCase__ ( cls : Optional[Any] ) -> int: __UpperCAmelCase : Optional[Any] = TOKEN HfFolder.save_token(snake_case ) @classmethod def lowerCamelCase__ ( cls : int ) -> Any: try: delete_repo(token=cls._token , repo_id='''test-feature-extractor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-feature-extractor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-feature-extractor''' ) except HTTPError: pass def lowerCamelCase__ ( self : List[str] ) -> Tuple: __UpperCAmelCase : Optional[Any] = WavaVecaFeatureExtractor.from_pretrained(snake_case ) feature_extractor.push_to_hub('''test-feature-extractor''' , use_auth_token=self._token ) __UpperCAmelCase : Optional[int] = WavaVecaFeatureExtractor.from_pretrained(f'{USER}/test-feature-extractor' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(snake_case , getattr(snake_case , snake_case ) ) # Reset repo delete_repo(token=self._token , repo_id='''test-feature-extractor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( snake_case , repo_id='''test-feature-extractor''' , push_to_hub=snake_case , use_auth_token=self._token ) __UpperCAmelCase : List[str] = WavaVecaFeatureExtractor.from_pretrained(f'{USER}/test-feature-extractor' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(snake_case , getattr(snake_case , snake_case ) ) def lowerCamelCase__ ( self : Optional[int] ) -> Dict: __UpperCAmelCase : Union[str, Any] = WavaVecaFeatureExtractor.from_pretrained(snake_case ) feature_extractor.push_to_hub('''valid_org/test-feature-extractor''' , use_auth_token=self._token ) __UpperCAmelCase : str = WavaVecaFeatureExtractor.from_pretrained('''valid_org/test-feature-extractor''' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(snake_case , getattr(snake_case , snake_case ) ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-feature-extractor''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( snake_case , repo_id='''valid_org/test-feature-extractor-org''' , push_to_hub=snake_case , use_auth_token=self._token ) __UpperCAmelCase : List[str] = WavaVecaFeatureExtractor.from_pretrained('''valid_org/test-feature-extractor-org''' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(snake_case , getattr(snake_case , snake_case ) ) def lowerCamelCase__ ( self : List[str] ) -> str: CustomFeatureExtractor.register_for_auto_class() __UpperCAmelCase : Dict = CustomFeatureExtractor.from_pretrained(snake_case ) feature_extractor.push_to_hub('''test-dynamic-feature-extractor''' , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( feature_extractor.auto_map , {'''AutoFeatureExtractor''': '''custom_feature_extraction.CustomFeatureExtractor'''} , ) __UpperCAmelCase : str = AutoFeatureExtractor.from_pretrained( f'{USER}/test-dynamic-feature-extractor' , trust_remote_code=snake_case ) # Can't make an isinstance check because the new_feature_extractor is from the CustomFeatureExtractor class of a dynamic module self.assertEqual(new_feature_extractor.__class__.__name__ , '''CustomFeatureExtractor''' )
240
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bert import BertTokenizer __UpperCAmelCase :Dict = logging.get_logger(__name__) __UpperCAmelCase :str = {"vocab_file": "vocab.txt", "tokenizer_file": "tokenizer.json"} __UpperCAmelCase :Optional[int] = { "vocab_file": { "bert-base-uncased": "https://huggingface.co/bert-base-uncased/resolve/main/vocab.txt", "bert-large-uncased": "https://huggingface.co/bert-large-uncased/resolve/main/vocab.txt", "bert-base-cased": "https://huggingface.co/bert-base-cased/resolve/main/vocab.txt", "bert-large-cased": "https://huggingface.co/bert-large-cased/resolve/main/vocab.txt", "bert-base-multilingual-uncased": ( "https://huggingface.co/bert-base-multilingual-uncased/resolve/main/vocab.txt" ), "bert-base-multilingual-cased": "https://huggingface.co/bert-base-multilingual-cased/resolve/main/vocab.txt", "bert-base-chinese": "https://huggingface.co/bert-base-chinese/resolve/main/vocab.txt", "bert-base-german-cased": "https://huggingface.co/bert-base-german-cased/resolve/main/vocab.txt", "bert-large-uncased-whole-word-masking": ( "https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/vocab.txt" ), "bert-large-cased-whole-word-masking": ( "https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/vocab.txt" ), "bert-large-uncased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt" ), "bert-large-cased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/vocab.txt" ), "bert-base-cased-finetuned-mrpc": ( "https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/vocab.txt" ), "bert-base-german-dbmdz-cased": "https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/vocab.txt", "bert-base-german-dbmdz-uncased": ( "https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/vocab.txt" ), "TurkuNLP/bert-base-finnish-cased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/vocab.txt" ), "TurkuNLP/bert-base-finnish-uncased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/vocab.txt" ), "wietsedv/bert-base-dutch-cased": ( "https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/vocab.txt" ), }, "tokenizer_file": { "bert-base-uncased": "https://huggingface.co/bert-base-uncased/resolve/main/tokenizer.json", "bert-large-uncased": "https://huggingface.co/bert-large-uncased/resolve/main/tokenizer.json", "bert-base-cased": "https://huggingface.co/bert-base-cased/resolve/main/tokenizer.json", "bert-large-cased": "https://huggingface.co/bert-large-cased/resolve/main/tokenizer.json", "bert-base-multilingual-uncased": ( "https://huggingface.co/bert-base-multilingual-uncased/resolve/main/tokenizer.json" ), "bert-base-multilingual-cased": ( "https://huggingface.co/bert-base-multilingual-cased/resolve/main/tokenizer.json" ), "bert-base-chinese": "https://huggingface.co/bert-base-chinese/resolve/main/tokenizer.json", "bert-base-german-cased": "https://huggingface.co/bert-base-german-cased/resolve/main/tokenizer.json", "bert-large-uncased-whole-word-masking": ( "https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/tokenizer.json" ), "bert-large-cased-whole-word-masking": ( "https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/tokenizer.json" ), "bert-large-uncased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json" ), "bert-large-cased-whole-word-masking-finetuned-squad": ( "https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/tokenizer.json" ), "bert-base-cased-finetuned-mrpc": ( "https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/tokenizer.json" ), "bert-base-german-dbmdz-cased": ( "https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/tokenizer.json" ), "bert-base-german-dbmdz-uncased": ( "https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/tokenizer.json" ), "TurkuNLP/bert-base-finnish-cased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/tokenizer.json" ), "TurkuNLP/bert-base-finnish-uncased-v1": ( "https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/tokenizer.json" ), "wietsedv/bert-base-dutch-cased": ( "https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/tokenizer.json" ), }, } __UpperCAmelCase :Union[str, Any] = { "bert-base-uncased": 5_1_2, "bert-large-uncased": 5_1_2, "bert-base-cased": 5_1_2, "bert-large-cased": 5_1_2, "bert-base-multilingual-uncased": 5_1_2, "bert-base-multilingual-cased": 5_1_2, "bert-base-chinese": 5_1_2, "bert-base-german-cased": 5_1_2, "bert-large-uncased-whole-word-masking": 5_1_2, "bert-large-cased-whole-word-masking": 5_1_2, "bert-large-uncased-whole-word-masking-finetuned-squad": 5_1_2, "bert-large-cased-whole-word-masking-finetuned-squad": 5_1_2, "bert-base-cased-finetuned-mrpc": 5_1_2, "bert-base-german-dbmdz-cased": 5_1_2, "bert-base-german-dbmdz-uncased": 5_1_2, "TurkuNLP/bert-base-finnish-cased-v1": 5_1_2, "TurkuNLP/bert-base-finnish-uncased-v1": 5_1_2, "wietsedv/bert-base-dutch-cased": 5_1_2, } __UpperCAmelCase :str = { "bert-base-uncased": {"do_lower_case": True}, "bert-large-uncased": {"do_lower_case": True}, "bert-base-cased": {"do_lower_case": False}, "bert-large-cased": {"do_lower_case": False}, "bert-base-multilingual-uncased": {"do_lower_case": True}, "bert-base-multilingual-cased": {"do_lower_case": False}, "bert-base-chinese": {"do_lower_case": False}, "bert-base-german-cased": {"do_lower_case": False}, "bert-large-uncased-whole-word-masking": {"do_lower_case": True}, "bert-large-cased-whole-word-masking": {"do_lower_case": False}, "bert-large-uncased-whole-word-masking-finetuned-squad": {"do_lower_case": True}, "bert-large-cased-whole-word-masking-finetuned-squad": {"do_lower_case": False}, "bert-base-cased-finetuned-mrpc": {"do_lower_case": False}, "bert-base-german-dbmdz-cased": {"do_lower_case": False}, "bert-base-german-dbmdz-uncased": {"do_lower_case": True}, "TurkuNLP/bert-base-finnish-cased-v1": {"do_lower_case": False}, "TurkuNLP/bert-base-finnish-uncased-v1": {"do_lower_case": True}, "wietsedv/bert-base-dutch-cased": {"do_lower_case": False}, } class a ( _a ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : List[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Tuple = PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : Optional[Any] = BertTokenizer def __init__( self : Optional[int] , snake_case : Union[str, Any]=None , snake_case : str=None , snake_case : Any=True , snake_case : Tuple="[UNK]" , snake_case : int="[SEP]" , snake_case : Optional[Any]="[PAD]" , snake_case : int="[CLS]" , snake_case : Optional[Any]="[MASK]" , snake_case : Union[str, Any]=True , snake_case : List[Any]=None , **snake_case : int , ) -> str: super().__init__( snake_case , tokenizer_file=snake_case , do_lower_case=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , tokenize_chinese_chars=snake_case , strip_accents=snake_case , **snake_case , ) __UpperCAmelCase : List[str] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , snake_case ) != do_lower_case or normalizer_state.get('''strip_accents''' , snake_case ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , snake_case ) != tokenize_chinese_chars ): __UpperCAmelCase : List[Any] = getattr(snake_case , normalizer_state.pop('''type''' ) ) __UpperCAmelCase : List[Any] = do_lower_case __UpperCAmelCase : List[Any] = strip_accents __UpperCAmelCase : str = tokenize_chinese_chars __UpperCAmelCase : List[str] = normalizer_class(**snake_case ) __UpperCAmelCase : Optional[int] = do_lower_case def lowerCamelCase__ ( self : List[Any] , snake_case : Tuple , snake_case : List[str]=None ) -> str: __UpperCAmelCase : Optional[int] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCamelCase__ ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ) -> List[int]: __UpperCAmelCase : List[str] = [self.sep_token_id] __UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase__ ( self : List[Any] , snake_case : str , snake_case : Optional[str] = None ) -> Tuple[str]: __UpperCAmelCase : Optional[int] = self._tokenizer.model.save(snake_case , name=snake_case ) return tuple(snake_case )
240
1
import socket def a_ ( ): '''simple docstring''' _lowerCamelCase : Optional[Any] =socket.socket(socket.AF_INET , socket.SOCK_STREAM ) _lowerCamelCase : Union[str, Any] =socket.gethostname() _lowerCamelCase : int =12_312 sock.connect((host, port) ) sock.send(b'Hello server!' ) with open('Received_file' , 'wb' ) as out_file: print('File opened' ) print('Receiving data...' ) while True: _lowerCamelCase : Optional[int] =sock.recv(1_024 ) if not data: break out_file.write(SCREAMING_SNAKE_CASE__ ) print('Successfully received the file' ) sock.close() print('Connection closed' ) if __name__ == "__main__": main()
199
import os import pytest from attr import dataclass lowerCamelCase = 'us-east-1' # defaults region @dataclass class A : UpperCamelCase__ : str UpperCamelCase__ : Dict ='arn:aws:iam::558105141721:role/sagemaker_execution_role' UpperCamelCase__ : int ={ 'task_name': 'mnli', 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'do_train': True, 'do_eval': True, 'do_predict': True, 'output_dir': '/opt/ml/model', 'overwrite_output_dir': True, 'max_steps': 500, 'save_steps': 5500, } UpperCamelCase__ : Optional[Any] ={**hyperparameters, 'max_steps': 1000} @property def lowerCamelCase ( self : List[Any] ) -> str: """simple docstring""" if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def lowerCamelCase ( self : List[Any] ) -> str: """simple docstring""" return F'''{self.framework}-transfromers-test''' @property def lowerCamelCase ( self : Union[str, Any] ) -> str: """simple docstring""" return F'''./tests/sagemaker/scripts/{self.framework}''' @property def lowerCamelCase ( self : Union[str, Any] ) -> str: """simple docstring""" if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='class' ) def a_ ( SCREAMING_SNAKE_CASE__ : Tuple ): '''simple docstring''' _lowerCamelCase : List[Any] =SageMakerTestEnvironment(framework=request.cls.framework )
199
1
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case__ : List[Any] = logging.get_logger(__name__) snake_case__ : Dict = { 'microsoft/biogpt': 'https://huggingface.co/microsoft/biogpt/resolve/main/config.json', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class A_ ( _lowerCamelCase ): lowerCAmelCase__ = """biogpt""" def __init__(self :str , _UpperCamelCase :int=4_2384 , _UpperCamelCase :Optional[Any]=1024 , _UpperCamelCase :List[str]=24 , _UpperCamelCase :str=16 , _UpperCamelCase :Union[str, Any]=4096 , _UpperCamelCase :List[Any]="gelu" , _UpperCamelCase :Tuple=0.1 , _UpperCamelCase :Tuple=0.1 , _UpperCamelCase :List[str]=1024 , _UpperCamelCase :Any=0.0_2 , _UpperCamelCase :List[str]=1e-12 , _UpperCamelCase :Optional[Any]=True , _UpperCamelCase :Dict=True , _UpperCamelCase :Union[str, Any]=0.0 , _UpperCamelCase :int=0.0 , _UpperCamelCase :Tuple=1 , _UpperCamelCase :Optional[Any]=0 , _UpperCamelCase :Any=2 , **_UpperCamelCase :Any , )-> Union[str, Any]: __A = vocab_size __A = max_position_embeddings __A = hidden_size __A = num_hidden_layers __A = num_attention_heads __A = intermediate_size __A = hidden_act __A = hidden_dropout_prob __A = attention_probs_dropout_prob __A = initializer_range __A = layer_norm_eps __A = scale_embedding __A = use_cache __A = layerdrop __A = activation_dropout super().__init__(pad_token_id=_UpperCamelCase , bos_token_id=_UpperCamelCase , eos_token_id=_UpperCamelCase , **_UpperCamelCase )
368
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) snake_case__ : Tuple = { 'configuration_layoutlmv2': ['LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LayoutLMv2Config'], 'processing_layoutlmv2': ['LayoutLMv2Processor'], 'tokenization_layoutlmv2': ['LayoutLMv2Tokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ : Union[str, Any] = ['LayoutLMv2TokenizerFast'] try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ : Optional[int] = ['LayoutLMv2FeatureExtractor'] snake_case__ : Dict = ['LayoutLMv2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: snake_case__ : Tuple = [ 'LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST', 'LayoutLMv2ForQuestionAnswering', 'LayoutLMv2ForSequenceClassification', 'LayoutLMv2ForTokenClassification', 'LayoutLMv2Layer', 'LayoutLMv2Model', 'LayoutLMv2PreTrainedModel', ] if TYPE_CHECKING: from .configuration_layoutlmva import LAYOUTLMV2_PRETRAINED_CONFIG_ARCHIVE_MAP, LayoutLMvaConfig from .processing_layoutlmva import LayoutLMvaProcessor from .tokenization_layoutlmva import LayoutLMvaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutlmva_fast import LayoutLMvaTokenizerFast try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_layoutlmva import LayoutLMvaFeatureExtractor, LayoutLMvaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_layoutlmva import ( LAYOUTLMV2_PRETRAINED_MODEL_ARCHIVE_LIST, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaLayer, LayoutLMvaModel, LayoutLMvaPreTrainedModel, ) else: import sys snake_case__ : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
250
0
import unittest from transformers import MobileBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertModel, ) class __lowerCAmelCase : def __init__(self , __magic_name__ , __magic_name__=13 , __magic_name__=7 , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=True , __magic_name__=99 , __magic_name__=64 , __magic_name__=32 , __magic_name__=5 , __magic_name__=4 , __magic_name__=37 , __magic_name__="gelu" , __magic_name__=0.1 , __magic_name__=0.1 , __magic_name__=512 , __magic_name__=16 , __magic_name__=2 , __magic_name__=0.02 , __magic_name__=3 , __magic_name__=4 , __magic_name__=None , ) -> str: '''simple docstring''' snake_case_ : List[str] = parent snake_case_ : Tuple = batch_size snake_case_ : str = seq_length snake_case_ : Optional[Any] = is_training snake_case_ : Optional[Any] = use_input_mask snake_case_ : Optional[int] = use_token_type_ids snake_case_ : Optional[Any] = use_labels snake_case_ : List[str] = vocab_size snake_case_ : Union[str, Any] = hidden_size snake_case_ : Tuple = embedding_size snake_case_ : Optional[int] = num_hidden_layers snake_case_ : List[str] = num_attention_heads snake_case_ : Optional[Any] = intermediate_size snake_case_ : Optional[Any] = hidden_act snake_case_ : Optional[int] = hidden_dropout_prob snake_case_ : Optional[int] = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : List[str] = type_vocab_size snake_case_ : int = type_sequence_label_size snake_case_ : Tuple = initializer_range snake_case_ : Tuple = num_labels snake_case_ : Tuple = num_choices snake_case_ : Tuple = scope def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ : Dict = None if self.use_input_mask: snake_case_ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ : Dict = None if self.use_token_type_ids: snake_case_ : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ : List[Any] = None snake_case_ : Dict = None snake_case_ : Tuple = None if self.use_labels: snake_case_ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase (self ) -> str: '''simple docstring''' return MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__magic_name__ , initializer_range=self.initializer_range , ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Dict: '''simple docstring''' snake_case_ : List[str] = MobileBertModel(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : Tuple = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ ) snake_case_ : Optional[int] = model(__magic_name__ , token_type_ids=__magic_name__ ) snake_case_ : Optional[Any] = model(__magic_name__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> int: '''simple docstring''' snake_case_ : Any = MobileBertForMaskedLM(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : Dict = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> List[str]: '''simple docstring''' snake_case_ : int = MobileBertForNextSentencePrediction(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : str = model( __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> str: '''simple docstring''' snake_case_ : Any = MobileBertForPreTraining(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : List[str] = model( __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ , next_sentence_label=__magic_name__ , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> List[str]: '''simple docstring''' snake_case_ : Tuple = MobileBertForQuestionAnswering(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : Any = model( __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , start_positions=__magic_name__ , end_positions=__magic_name__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> str: '''simple docstring''' snake_case_ : Optional[int] = self.num_labels snake_case_ : Tuple = MobileBertForSequenceClassification(__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : List[Any] = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Optional[Any]: '''simple docstring''' snake_case_ : int = self.num_labels snake_case_ : Dict = MobileBertForTokenClassification(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : List[str] = model(__magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> Any: '''simple docstring''' snake_case_ : Tuple = self.num_choices snake_case_ : str = MobileBertForMultipleChoice(config=__magic_name__ ) model.to(__magic_name__ ) model.eval() snake_case_ : Dict = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ : Tuple = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ : int = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ : List[Any] = model( __magic_name__ , attention_mask=__magic_name__ , token_type_ids=__magic_name__ , labels=__magic_name__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' snake_case_ : Optional[int] = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) : List[Any] = config_and_inputs snake_case_ : List[str] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __lowerCAmelCase ( _a, _a, unittest.TestCase ): lowerCamelCase_ : int = ( ( MobileBertModel, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase_ : Union[str, Any] = ( { '''feature-extraction''': MobileBertModel, '''fill-mask''': MobileBertForMaskedLM, '''question-answering''': MobileBertForQuestionAnswering, '''text-classification''': MobileBertForSequenceClassification, '''token-classification''': MobileBertForTokenClassification, '''zero-shot''': MobileBertForSequenceClassification, } if is_torch_available() else {} ) lowerCamelCase_ : Union[str, Any] = True def lowerCamelCase (self , __magic_name__ , __magic_name__ , __magic_name__=False ) -> List[str]: '''simple docstring''' snake_case_ : Dict = super()._prepare_for_class(__magic_name__ , __magic_name__ , return_labels=__magic_name__ ) if return_labels: if model_class in get_values(__magic_name__ ): snake_case_ : Tuple = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=__magic_name__ ) snake_case_ : int = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=__magic_name__ ) return inputs_dict def lowerCamelCase (self ) -> List[str]: '''simple docstring''' snake_case_ : Tuple = MobileBertModelTester(self ) snake_case_ : List[str] = ConfigTester(self , config_class=__magic_name__ , hidden_size=37 ) def lowerCamelCase (self ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase (self ) -> Tuple: '''simple docstring''' snake_case_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*__magic_name__ ) def lowerCamelCase (self ) -> Any: '''simple docstring''' snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*__magic_name__ ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*__magic_name__ ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*__magic_name__ ) def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*__magic_name__ ) def lowerCamelCase (self ) -> str: '''simple docstring''' snake_case_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*__magic_name__ ) def lowerCamelCase (self ) -> Dict: '''simple docstring''' snake_case_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*__magic_name__ ) def lowerCamelCase (self ) -> List[Any]: '''simple docstring''' snake_case_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*__magic_name__ ) def lowerCamelCase_ ( _UpperCamelCase ) -> Optional[int]: """simple docstring""" return torch.tensor( _UpperCamelCase , dtype=torch.long , device=_UpperCamelCase , ) lowerCAmelCase_ = 1e-3 @require_torch @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase (self ) -> Optional[Any]: '''simple docstring''' snake_case_ : Optional[Any] = MobileBertModel.from_pretrained('''google/mobilebert-uncased''' ).to(__magic_name__ ) snake_case_ : Optional[Any] = _long_tensor([[101, 7110, 1005, 1056, 2023, 1_1333, 1_7413, 1029, 102]] ) with torch.no_grad(): snake_case_ : Optional[int] = model(__magic_name__ )[0] snake_case_ : Union[str, Any] = torch.Size((1, 9, 512) ) self.assertEqual(output.shape , __magic_name__ ) snake_case_ : List[str] = torch.tensor( [ [ [-2.473_6526e07, 8.269_1656e04, 1.652_1838e05], [-5.754_1704e-01, 3.905_6022e00, 4.401_1507e00], [2.604_7359e00, 1.567_7652e00, -1.732_4188e-01], ] ] , device=__magic_name__ , ) # MobileBERT results range from 10e0 to 10e8. Even a 0.0000001% difference with a value of 10e8 results in a # ~1 difference, it's therefore not a good idea to measure using addition. # Here, we instead divide the expected result with the result in order to obtain ~1. We then check that the # result is held between bounds: 1 - TOLERANCE < expected_result / result < 1 + TOLERANCE snake_case_ : Any = torch.all((expected_slice / output[..., :3, :3]) >= 1 - TOLERANCE ) snake_case_ : Union[str, Any] = torch.all((expected_slice / output[..., :3, :3]) <= 1 + TOLERANCE ) self.assertTrue(lower_bound and upper_bound )
279
from math import factorial lowerCAmelCase_ = {str(digit): factorial(digit) for digit in range(1_0)} def lowerCamelCase_ ( _UpperCamelCase ) -> int: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ): raise TypeError('''Parameter number must be int''' ) if number < 0: raise ValueError('''Parameter number must be greater than or equal to 0''' ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(_UpperCamelCase ) ) def lowerCamelCase_ ( _UpperCamelCase = 60 , _UpperCamelCase = 1_000_000 ) -> int: """simple docstring""" if not isinstance(_UpperCamelCase , _UpperCamelCase ) or not isinstance(_UpperCamelCase , _UpperCamelCase ): raise TypeError('''Parameters chain_length and number_limit must be int''' ) if chain_length <= 0 or number_limit <= 0: raise ValueError( '''Parameters chain_length and number_limit must be greater than 0''' ) # the counter for the chains with the exact desired length snake_case_ : Optional[Any] = 0 # the cached sizes of the previous chains snake_case_ : dict[int, int] = {} for start_chain_element in range(1 , _UpperCamelCase ): # The temporary set will contain the elements of the chain snake_case_ : List[str] = set() snake_case_ : List[Any] = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. snake_case_ : Any = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(_UpperCamelCase ) chain_set_length += 1 snake_case_ : List[Any] = digit_factorial_sum(_UpperCamelCase ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] snake_case_ : List[str] = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F'''{solution()}''')
279
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase : Dict = {"configuration_xlnet": ["XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLNetConfig"]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[str] = ["XLNetTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : List[Any] = ["XLNetTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : int = [ "XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "XLNetForMultipleChoice", "XLNetForQuestionAnswering", "XLNetForQuestionAnsweringSimple", "XLNetForSequenceClassification", "XLNetForTokenClassification", "XLNetLMHeadModel", "XLNetModel", "XLNetPreTrainedModel", "load_tf_weights_in_xlnet", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : Union[str, Any] = [ "TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLNetForMultipleChoice", "TFXLNetForQuestionAnsweringSimple", "TFXLNetForSequenceClassification", "TFXLNetForTokenClassification", "TFXLNetLMHeadModel", "TFXLNetMainLayer", "TFXLNetModel", "TFXLNetPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys lowerCamelCase : List[str] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
208
import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() lowerCamelCase : Union[str, Any] = logging.get_logger(__name__) lowerCamelCase : str = {name: getattr(transformers, name + "Fast") for name in SLOW_TO_FAST_CONVERTERS} def _SCREAMING_SNAKE_CASE ( lowercase : Any , lowercase : str , lowercase : int , lowercase : List[Any] ): '''simple docstring''' if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(f"""Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.""" ) if tokenizer_name is None: lowerCamelCase_ = TOKENIZER_CLASSES else: lowerCamelCase_ = {tokenizer_name: getattr(lowercase , tokenizer_name + 'Fast' )} logger.info(f"""Loading tokenizer classes: {tokenizer_names}""" ) for tokenizer_name in tokenizer_names: lowerCamelCase_ = TOKENIZER_CLASSES[tokenizer_name] lowerCamelCase_ = True if checkpoint_name is None: lowerCamelCase_ = list(tokenizer_class.max_model_input_sizes.keys() ) else: lowerCamelCase_ = [checkpoint_name] logger.info(f"""For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}""" ) for checkpoint in checkpoint_names: logger.info(f"""Loading {tokenizer_class.__class__.__name__} {checkpoint}""" ) # Load tokenizer lowerCamelCase_ = tokenizer_class.from_pretrained(lowercase , force_download=lowercase ) # Save fast tokenizer logger.info(f"""Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}""" ) # For organization names we create sub-directories if "/" in checkpoint: lowerCamelCase_ , lowerCamelCase_ = checkpoint.split('/' ) lowerCamelCase_ = os.path.join(lowercase , lowercase ) elif add_prefix: lowerCamelCase_ = checkpoint lowerCamelCase_ = dump_path else: lowerCamelCase_ = None lowerCamelCase_ = dump_path logger.info(f"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: lowerCamelCase_ = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] lowerCamelCase_ = file_path.split(lowercase )[-1][0] if next_char == "/": lowerCamelCase_ = os.path.join(lowercase , lowercase ) lowerCamelCase_ = None logger.info(f"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) lowerCamelCase_ = tokenizer.save_pretrained( lowercase , legacy_format=lowercase , filename_prefix=lowercase ) logger.info(f"""=> File names {file_names}""" ) for file_name in file_names: if not file_name.endswith('tokenizer.json' ): os.remove(lowercase ) logger.info(f"""=> removing {file_name}""" ) if __name__ == "__main__": lowerCamelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--dump_path", default=None, type=str, required=True, help="Path to output generated fast tokenizer files." ) parser.add_argument( "--tokenizer_name", default=None, type=str, help=( F"""Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will """ "download and convert all the checkpoints from AWS." ), ) parser.add_argument( "--checkpoint_name", default=None, type=str, help="Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.", ) parser.add_argument( "--force_download", action="store_true", help="Re-download checkpoints.", ) lowerCamelCase : Union[str, Any] = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
208
1
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class _UpperCAmelCase ( lowercase_ , lowercase_ , lowercase_ ): UpperCamelCase = [R'''h\.\d+\.attn\.bias''', R'''h\.\d+\.attn\.masked_bias'''] @register_to_config def __init__( self :Union[str, Any] , __UpperCamelCase :int , __UpperCamelCase :int , __UpperCamelCase :Optional[int] = None , __UpperCamelCase :int = 5_02_57 , __UpperCamelCase :int = 10_24 , __UpperCamelCase :int = 7_68 , __UpperCamelCase :int = 12 , __UpperCamelCase :int = 12 , __UpperCamelCase :Optional[int] = None , __UpperCamelCase :str = "gelu_new" , __UpperCamelCase :float = 0.1 , __UpperCamelCase :float = 0.1 , __UpperCamelCase :float = 0.1 , __UpperCamelCase :float = 1e-5 , __UpperCamelCase :float = 0.02 , __UpperCamelCase :bool = True , __UpperCamelCase :bool = True , __UpperCamelCase :bool = False , __UpperCamelCase :bool = False , ): super().__init__() A = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( f"`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and" f" `n_embd`: {n_embd} are not equal." ) A = prefix_inner_dim A = prefix_hidden_dim A = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) A = ( nn.Linear(self.prefix_hidden_dim , __UpperCamelCase ) if self.prefix_hidden_dim is not None else nn.Identity() ) A = GPTaConfig( vocab_size=__UpperCamelCase , n_positions=__UpperCamelCase , n_embd=__UpperCamelCase , n_layer=__UpperCamelCase , n_head=__UpperCamelCase , n_inner=__UpperCamelCase , activation_function=__UpperCamelCase , resid_pdrop=__UpperCamelCase , embd_pdrop=__UpperCamelCase , attn_pdrop=__UpperCamelCase , layer_norm_epsilon=__UpperCamelCase , initializer_range=__UpperCamelCase , scale_attn_weights=__UpperCamelCase , use_cache=__UpperCamelCase , scale_attn_by_inverse_layer_idx=__UpperCamelCase , reorder_and_upcast_attn=__UpperCamelCase , ) A = GPTaLMHeadModel(__UpperCamelCase ) def lowerCamelCase ( self :int , __UpperCamelCase :torch.Tensor , __UpperCamelCase :torch.Tensor , __UpperCamelCase :Optional[torch.Tensor] = None , __UpperCamelCase :Optional[torch.Tensor] = None , ): A = self.transformer.transformer.wte(__UpperCamelCase ) A = self.encode_prefix(__UpperCamelCase ) A = self.decode_prefix(__UpperCamelCase ) A = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: A = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) A = torch.cat((dummy_token, input_ids) , dim=1 ) A = self.transformer(inputs_embeds=__UpperCamelCase , labels=__UpperCamelCase , attention_mask=__UpperCamelCase ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def lowerCamelCase ( self :Optional[Any] , __UpperCamelCase :int , __UpperCamelCase :torch.device ): return torch.zeros(__UpperCamelCase , self.prefix_length , dtype=torch.intaa , device=__UpperCamelCase ) def lowerCamelCase ( self :List[str] , __UpperCamelCase :Any ): return self.encode_prefix(__UpperCamelCase ) @torch.no_grad() def lowerCamelCase ( self :Dict , __UpperCamelCase :str , __UpperCamelCase :Union[str, Any] , __UpperCamelCase :List[Any] ): A = torch.split(__UpperCamelCase , 1 , dim=0 ) A = [] A = [] for feature in features: A = self.decode_prefix(feature.to(__UpperCamelCase ) ) # back to the clip feature # Only support beam search for now A, A = self.generate_beam( input_embeds=__UpperCamelCase , device=__UpperCamelCase , eos_token_id=__UpperCamelCase ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) A = torch.stack(__UpperCamelCase ) A = torch.stack(__UpperCamelCase ) return generated_tokens, generated_seq_lengths @torch.no_grad() def lowerCamelCase ( self :List[str] , __UpperCamelCase :Optional[int]=None , __UpperCamelCase :Any=None , __UpperCamelCase :Optional[int]=None , __UpperCamelCase :int = 5 , __UpperCamelCase :int = 67 , __UpperCamelCase :float = 1.0 , __UpperCamelCase :Optional[int] = None , ): A = eos_token_id A = None A = None A = torch.ones(__UpperCamelCase , device=__UpperCamelCase , dtype=torch.int ) A = torch.zeros(__UpperCamelCase , device=__UpperCamelCase , dtype=torch.bool ) if input_embeds is not None: A = input_embeds else: A = self.transformer.transformer.wte(__UpperCamelCase ) for i in range(__UpperCamelCase ): A = self.transformer(inputs_embeds=__UpperCamelCase ) A = outputs.logits A = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) A = logits.softmax(-1 ).log() if scores is None: A, A = logits.topk(__UpperCamelCase , -1 ) A = generated.expand(__UpperCamelCase , *generated.shape[1:] ) A, A = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: A = next_tokens else: A = tokens.expand(__UpperCamelCase , *tokens.shape[1:] ) A = torch.cat((tokens, next_tokens) , dim=1 ) else: A = -float(np.inf ) A = 0 A = scores[:, None] + logits seq_lengths[~is_stopped] += 1 A = scores_sum / seq_lengths[:, None] A, A = scores_sum_average.view(-1 ).topk(__UpperCamelCase , -1 ) A = next_tokens // scores_sum.shape[1] A = seq_lengths[next_tokens_source] A = next_tokens % scores_sum.shape[1] A = next_tokens.unsqueeze(1 ) A = tokens[next_tokens_source] A = torch.cat((tokens, next_tokens) , dim=1 ) A = generated[next_tokens_source] A = scores_sum_average * seq_lengths A = is_stopped[next_tokens_source] A = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) A = torch.cat((generated, next_token_embed) , dim=1 ) A = is_stopped + next_tokens.eq(__UpperCamelCase ).squeeze() if is_stopped.all(): break A = scores / seq_lengths A = scores.argsort(descending=__UpperCamelCase ) # tokens tensors are already padded to max_seq_length A = [tokens[i] for i in order] A = torch.stack(__UpperCamelCase , dim=0 ) A = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
292
"""simple docstring""" import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class _UpperCAmelCase : UpperCamelCase = None def lowerCamelCase ( self :List[Any] ): A = self.feature_extraction_class(**self.feat_extract_dict ) A = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] , __UpperCamelCase ) def lowerCamelCase ( self :Dict ): A = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A = os.path.join(__UpperCamelCase , "feat_extract.json" ) feat_extract_first.to_json_file(__UpperCamelCase ) A = self.feature_extraction_class.from_json_file(__UpperCamelCase ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def lowerCamelCase ( self :Dict ): A = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A = feat_extract_first.save_pretrained(__UpperCamelCase )[0] check_json_file_has_correct_format(__UpperCamelCase ) A = self.feature_extraction_class.from_pretrained(__UpperCamelCase ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def lowerCamelCase ( self :Tuple ): A = self.feature_extraction_class() self.assertIsNotNone(__UpperCamelCase )
292
1
"""simple docstring""" import os import re from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a = logging.get_logger(__name__) a = { '''vocab_file''': '''vocab.txt''', '''merges_file''': '''bpe.codes''', } a = { '''vocab_file''': { '''vinai/phobert-base''': '''https://huggingface.co/vinai/phobert-base/resolve/main/vocab.txt''', '''vinai/phobert-large''': '''https://huggingface.co/vinai/phobert-large/resolve/main/vocab.txt''', }, '''merges_file''': { '''vinai/phobert-base''': '''https://huggingface.co/vinai/phobert-base/resolve/main/bpe.codes''', '''vinai/phobert-large''': '''https://huggingface.co/vinai/phobert-large/resolve/main/bpe.codes''', }, } a = { '''vinai/phobert-base''': 256, '''vinai/phobert-large''': 256, } def _snake_case ( _snake_case : str ) -> Optional[Any]: '''simple docstring''' _A = set() _A = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _A = char _A = set(_snake_case ) return pairs class lowercase_ ( snake_case_ ): '''simple docstring''' UpperCAmelCase : Tuple = VOCAB_FILES_NAMES UpperCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : List[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Dict="<s>" , _UpperCAmelCase : Optional[Any]="</s>" , _UpperCAmelCase : Dict="</s>" , _UpperCAmelCase : Dict="<s>" , _UpperCAmelCase : Dict="<unk>" , _UpperCAmelCase : List[Any]="<pad>" , _UpperCAmelCase : Optional[Any]="<mask>" , **_UpperCAmelCase : str , ): super().__init__( bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , unk_token=_UpperCAmelCase , sep_token=_UpperCAmelCase , cls_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , mask_token=_UpperCAmelCase , **_UpperCAmelCase , ) _A = vocab_file _A = merges_file _A = {} _A = 0 _A = 1 _A = 2 _A = 3 self.add_from_file(_UpperCAmelCase ) _A = {v: k for k, v in self.encoder.items()} with open(_UpperCAmelCase , encoding='utf-8' ) as merges_handle: _A = merges_handle.read().split('\n' )[:-1] _A = [tuple(merge.split()[:-1] ) for merge in merges] _A = dict(zip(_UpperCAmelCase , range(len(_UpperCAmelCase ) ) ) ) _A = {} def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _A = [self.cls_token_id] _A = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None , _UpperCAmelCase : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_UpperCAmelCase , token_ids_a=_UpperCAmelCase , already_has_special_tokens=_UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(_UpperCAmelCase )) + [1] return [1] + ([0] * len(_UpperCAmelCase )) + [1, 1] + ([0] * len(_UpperCAmelCase )) + [1] def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : List[int] , _UpperCAmelCase : Optional[List[int]] = None ): _A = [self.sep_token_id] _A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def lowerCAmelCase_ ( self : str ): return len(self.encoder ) def lowerCAmelCase_ ( self : int ): return dict(self.encoder , **self.added_tokens_encoder ) def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : Tuple ): if token in self.cache: return self.cache[token] _A = tuple(_UpperCAmelCase ) _A = tuple(list(word[:-1] ) + [word[-1] + '</w>'] ) _A = get_pairs(_UpperCAmelCase ) if not pairs: return token while True: _A = min(_UpperCAmelCase , key=lambda _UpperCAmelCase : self.bpe_ranks.get(_UpperCAmelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break _A = bigram _A = [] _A = 0 while i < len(_UpperCAmelCase ): try: _A = word.index(_UpperCAmelCase , _UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _A = j if word[i] == first and i < len(_UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _A = tuple(_UpperCAmelCase ) _A = new_word if len(_UpperCAmelCase ) == 1: break else: _A = get_pairs(_UpperCAmelCase ) _A = '@@ '.join(_UpperCAmelCase ) _A = word[:-4] _A = word return word def lowerCAmelCase_ ( self : List[Any] , _UpperCAmelCase : Any ): _A = [] _A = re.findall(r'\S+\n?' , _UpperCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(_UpperCAmelCase ).split(' ' ) ) ) return split_tokens def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : Tuple ): return self.encoder.get(_UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowerCAmelCase_ ( self : str , _UpperCAmelCase : Any ): return self.decoder.get(_UpperCAmelCase , self.unk_token ) def lowerCAmelCase_ ( self : str , _UpperCAmelCase : List[Any] ): _A = ' '.join(_UpperCAmelCase ).replace('@@ ' , '' ).strip() return out_string def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : str , _UpperCAmelCase : Optional[str] = None ): if not os.path.isdir(_UpperCAmelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _A = os.path.join( _UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) _A = os.path.join( _UpperCAmelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCAmelCase ): copyfile(self.vocab_file , _UpperCAmelCase ) if os.path.abspath(self.merges_file ) != os.path.abspath(_UpperCAmelCase ): copyfile(self.merges_file , _UpperCAmelCase ) return out_vocab_file, out_merge_file def lowerCAmelCase_ ( self : List[str] , _UpperCAmelCase : Optional[int] ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): try: with open(_UpperCAmelCase , 'r' , encoding='utf-8' ) as fd: self.add_from_file(_UpperCAmelCase ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception(F'''Incorrect encoding detected in {f}, please rebuild the dataset''' ) return _A = f.readlines() for lineTmp in lines: _A = lineTmp.strip() _A = line.rfind(' ' ) if idx == -1: raise ValueError('Incorrect dictionary format, expected \'<token> <cnt>\'' ) _A = line[:idx] _A = len(self.encoder )
356
"""simple docstring""" from __future__ import annotations import time import numpy as np a = [8, 5, 9, 7] a = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] a = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1, 0, 5], [1, 5, 3, 0], [3, 0, 3, 3], ] class lowercase_ : '''simple docstring''' def __init__( self : str , _UpperCAmelCase : list[int] , _UpperCAmelCase : list[list[int]] , _UpperCAmelCase : list[list[int]] , ): _A = claim_vector _A = allocated_resources_table _A = maximum_claim_table def lowerCAmelCase_ ( self : Tuple ): return [ sum(p_item[i] for p_item in self.__allocated_resources_table ) for i in range(len(self.__allocated_resources_table[0] ) ) ] def lowerCAmelCase_ ( self : Tuple ): return np.array(self.__claim_vector ) - np.array( self.__processes_resource_summation() ) def lowerCAmelCase_ ( self : List[Any] ): return [ list(np.array(self.__maximum_claim_table[i] ) - np.array(_UpperCAmelCase ) ) for i, allocated_resource in enumerate(self.__allocated_resources_table ) ] def lowerCAmelCase_ ( self : List[Any] ): return {self.__need().index(_UpperCAmelCase ): i for i in self.__need()} def lowerCAmelCase_ ( self : List[str] , **_UpperCAmelCase : int ): _A = self.__need() _A = self.__allocated_resources_table _A = self.__available_resources() _A = self.__need_index_manager() for kw, val in kwargs.items(): if kw and val is True: self.__pretty_data() print('_' * 50 + '\n' ) while need_list: _A = False for each_need in need_list: _A = True for index, need in enumerate(_UpperCAmelCase ): if need > available_resources[index]: _A = False break if execution: _A = True # get the original index of the process from ind_ctrl db for original_need_index, need_clone in need_index_manager.items(): if each_need == need_clone: _A = original_need_index print(F'''Process {process_number + 1} is executing.''' ) # remove the process run from stack need_list.remove(_UpperCAmelCase ) # update available/freed resources stack _A = np.array(_UpperCAmelCase ) + np.array( alloc_resources_table[process_number] ) print( 'Updated available resource stack for processes: ' + ' '.join([str(_UpperCAmelCase ) for x in available_resources] ) ) break if safe: print('The process is in a safe state.\n' ) else: print('System in unsafe state. Aborting...\n' ) break def lowerCAmelCase_ ( self : Union[str, Any] ): print(' ' * 9 + 'Allocated Resource Table' ) for item in self.__allocated_resources_table: print( F'''P{self.__allocated_resources_table.index(_UpperCAmelCase ) + 1}''' + ' '.join(F'''{it:>8}''' for it in item ) + '\n' ) print(' ' * 9 + 'System Resource Table' ) for item in self.__maximum_claim_table: print( F'''P{self.__maximum_claim_table.index(_UpperCAmelCase ) + 1}''' + ' '.join(F'''{it:>8}''' for it in item ) + '\n' ) print( 'Current Usage by Active Processes: ' + ' '.join(str(_UpperCAmelCase ) for x in self.__claim_vector ) ) print( 'Initial Available Resources: ' + ' '.join(str(_UpperCAmelCase ) for x in self.__available_resources() ) ) time.sleep(1 ) if __name__ == "__main__": import doctest doctest.testmod()
271
0
'''simple docstring''' from graphs.minimum_spanning_tree_kruskal import kruskal def _lowerCAmelCase ( ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE =9 _SCREAMING_SNAKE_CASE =[ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] _SCREAMING_SNAKE_CASE =kruskal(_UpperCAmelCase , _UpperCAmelCase ) _SCREAMING_SNAKE_CASE =[ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] assert sorted(_UpperCAmelCase ) == sorted(_UpperCAmelCase )
47
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo UpperCAmelCase__ = "\\n@misc{wu2016googles,\n title={Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey\n and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin\n Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto\n Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and\n Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes\n and Jeffrey Dean},\n year={2016},\n eprint={1609.08144},\n archivePrefix={arXiv},\n primaryClass={cs.CL}\n}\n" UpperCAmelCase__ = "\\nThe BLEU score has some undesirable properties when used for single\nsentences, as it was designed to be a corpus measure. We therefore\nuse a slightly different score for our RL experiments which we call\nthe 'GLEU score'. For the GLEU score, we record all sub-sequences of\n1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then\ncompute a recall, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the target (ground truth) sequence,\nand a precision, which is the ratio of the number of matching n-grams\nto the number of total n-grams in the generated output sequence. Then\nGLEU score is simply the minimum of recall and precision. This GLEU\nscore's range is always between 0 (no matches) and 1 (all match) and\nit is symmetrical when switching output and target. According to\nour experiments, GLEU score correlates quite well with the BLEU\nmetric on a corpus level but does not have its drawbacks for our per\nsentence reward objective.\n" UpperCAmelCase__ = "\\nComputes corpus-level Google BLEU (GLEU) score of translated segments against one or more references.\nInstead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching\ntokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values.\n\nArgs:\n predictions (list of str): list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references (list of list of str): list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n min_len (int): The minimum order of n-gram this function should extract. Defaults to 1.\n max_len (int): The maximum order of n-gram this function should extract. Defaults to 4.\n\nReturns:\n 'google_bleu': google_bleu score\n\nExamples:\n Example 1:\n >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which',\n ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always',\n ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat']\n >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which',\n ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never',\n ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat']\n\n >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was',\n ... 'interested', 'in', 'world', 'history']\n >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history',\n ... 'because', 'he', 'read', 'the', 'book']\n\n >>> list_of_references = [[ref1a], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric(\"google_bleu\")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results[\"google_bleu\"], 2))\n 0.44\n\n Example 2:\n >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which',\n ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always',\n ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat']\n >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which',\n ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never',\n ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat']\n >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that',\n ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never',\n ... 'heed', 'the', 'cat', 'commands']\n >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the',\n ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions',\n ... 'of', 'the', 'cat']\n\n >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was',\n ... 'interested', 'in', 'world', 'history']\n >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history',\n ... 'because', 'he', 'read', 'the', 'book']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric(\"google_bleu\")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references)\n >>> print(round(results[\"google_bleu\"], 2))\n 0.61\n\n Example 3:\n >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which',\n ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always',\n ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat']\n >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which',\n ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never',\n ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat']\n >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that',\n ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never',\n ... 'heed', 'the', 'cat', 'commands']\n >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the',\n ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions',\n ... 'of', 'the', 'cat']\n\n >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was',\n ... 'interested', 'in', 'world', 'history']\n >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history',\n ... 'because', 'he', 'read', 'the', 'book']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric(\"google_bleu\")\n >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2)\n >>> print(round(results[\"google_bleu\"], 2))\n 0.53\n\n Example 4:\n >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which',\n ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always',\n ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat']\n >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which',\n ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never',\n ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat']\n >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that',\n ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never',\n ... 'heed', 'the', 'cat', 'commands']\n >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the',\n ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions',\n ... 'of', 'the', 'cat']\n\n >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was',\n ... 'interested', 'in', 'world', 'history']\n >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history',\n ... 'because', 'he', 'read', 'the', 'book']\n\n >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]]\n >>> hypotheses = [hyp1, hyp2]\n >>> google_bleu = datasets.load_metric(\"google_bleu\")\n >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6)\n >>> print(round(results[\"google_bleu\"], 2))\n 0.4\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): def _lowerCamelCase ( self : str) -> MetricInfo: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' , id='token') , id='sequence'), 'references': datasets.Sequence( datasets.Sequence(datasets.Value('string' , id='token') , id='sequence') , id='references'), }) , ) def _lowerCamelCase ( self : Union[str, Any] , A : List[List[List[str]]] , A : List[List[str]] , A : int = 1 , A : int = 4 , ) -> Dict[str, float]: """simple docstring""" return { "google_bleu": gleu_score.corpus_gleu( list_of_references=A , hypotheses=A , min_len=A , max_len=A) }
339
0
"""simple docstring""" from typing import List, Union import numpy as np from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, logging from .base import PIPELINE_INIT_ARGS, ArgumentHandler, ChunkPipeline __snake_case = logging.get_logger(__name__) class __lowerCamelCase ( a__ ): '''simple docstring''' def _UpperCAmelCase ( self , __UpperCAmelCase ) -> str: if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _a = [label.strip() for label in labels.split(''',''' ) if label.strip()] return labels def __call__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[int]: if len(__UpperCAmelCase ) == 0 or len(__UpperCAmelCase ) == 0: raise ValueError('''You must include at least one label and at least one sequence.''' ) if hypothesis_template.format(labels[0] ) == hypothesis_template: raise ValueError( ( '''The provided hypothesis_template "{}" was not able to be formatted with the target labels. ''' '''Make sure the passed template includes formatting syntax such as {{}} where the label should go.''' ).format(__UpperCAmelCase ) ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ): _a = [sequences] _a = [] for sequence in sequences: sequence_pairs.extend([[sequence, hypothesis_template.format(__UpperCAmelCase )] for label in labels] ) return sequence_pairs, sequences @add_end_docstrings(a__ ) class __lowerCamelCase ( a__ ): '''simple docstring''' def __init__( self , __UpperCAmelCase=ZeroShotClassificationArgumentHandler() , *__UpperCAmelCase , **__UpperCAmelCase ) -> int: _a = args_parser super().__init__(*__UpperCAmelCase , **__UpperCAmelCase ) if self.entailment_id == -1: logger.warning( '''Failed to determine \'entailment\' label id from the label2id mapping in the model config. Setting to ''' '''-1. Define a descriptive label2id mapping in the model config to ensure correct outputs.''' ) @property def _UpperCAmelCase ( self ) -> Optional[Any]: for label, ind in self.model.config.labelaid.items(): if label.lower().startswith('''entail''' ): return ind return -1 def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=TruncationStrategy.ONLY_FIRST , **__UpperCAmelCase ) -> Union[str, Any]: _a = self.framework if self.tokenizer.pad_token is None: # Override for tokenizers not supporting padding logger.error( '''Tokenizer was not supporting padding necessary for zero-shot, attempting to use ''' ''' `pad_token=eos_token`''' ) _a = self.tokenizer.eos_token try: _a = self.tokenizer( __UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors=__UpperCAmelCase , padding=__UpperCAmelCase , truncation=__UpperCAmelCase , ) except Exception as e: if "too short" in str(__UpperCAmelCase ): # tokenizers might yell that we want to truncate # to a value that is not even reached by the input. # In that case we don't want to truncate. # It seems there's not a really better way to catch that # exception. _a = self.tokenizer( __UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors=__UpperCAmelCase , padding=__UpperCAmelCase , truncation=TruncationStrategy.DO_NOT_TRUNCATE , ) else: raise e return inputs def _UpperCAmelCase ( self , **__UpperCAmelCase ) -> List[Any]: if kwargs.get('''multi_class''' , __UpperCAmelCase ) is not None: _a = kwargs['''multi_class'''] logger.warning( '''The `multi_class` argument has been deprecated and renamed to `multi_label`. ''' '''`multi_class` will be removed in a future version of Transformers.''' ) _a = {} if "candidate_labels" in kwargs: _a = self._args_parser._parse_labels(kwargs['''candidate_labels'''] ) if "hypothesis_template" in kwargs: _a = kwargs['''hypothesis_template'''] _a = {} if "multi_label" in kwargs: _a = kwargs['''multi_label'''] return preprocess_params, {}, postprocess_params def __call__( self , __UpperCAmelCase , *__UpperCAmelCase , **__UpperCAmelCase , ) -> List[str]: if len(__UpperCAmelCase ) == 0: pass elif len(__UpperCAmelCase ) == 1 and "candidate_labels" not in kwargs: _a = args[0] else: raise ValueError(F'Unable to understand extra arguments {args}' ) return super().__call__(__UpperCAmelCase , **__UpperCAmelCase ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase=None , __UpperCAmelCase="This example is {}." ) -> Optional[Any]: _a , _a = self._args_parser(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) for i, (candidate_label, sequence_pair) in enumerate(zip(__UpperCAmelCase , __UpperCAmelCase ) ): _a = self._parse_and_tokenize([sequence_pair] ) yield { "candidate_label": candidate_label, "sequence": sequences[0], "is_last": i == len(__UpperCAmelCase ) - 1, **model_input, } def _UpperCAmelCase ( self , __UpperCAmelCase ) -> Any: _a = inputs['''candidate_label'''] _a = inputs['''sequence'''] _a = {k: inputs[k] for k in self.tokenizer.model_input_names} _a = self.model(**__UpperCAmelCase ) _a = { '''candidate_label''': candidate_label, '''sequence''': sequence, '''is_last''': inputs['''is_last'''], **outputs, } return model_outputs def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase=False ) -> str: _a = [outputs['''candidate_label'''] for outputs in model_outputs] _a = [outputs['''sequence'''] for outputs in model_outputs] _a = np.concatenate([output['''logits'''].numpy() for output in model_outputs] ) _a = logits.shape[0] _a = len(__UpperCAmelCase ) _a = N // n _a = logits.reshape((num_sequences, n, -1) ) if multi_label or len(__UpperCAmelCase ) == 1: # softmax over the entailment vs. contradiction dim for each label independently _a = self.entailment_id _a = -1 if entailment_id == 0 else 0 _a = reshaped_outputs[..., [contradiction_id, entailment_id]] _a = np.exp(__UpperCAmelCase ) / np.exp(__UpperCAmelCase ).sum(-1 , keepdims=__UpperCAmelCase ) _a = scores[..., 1] else: # softmax the "entailment" logits over all candidate labels _a = reshaped_outputs[..., self.entailment_id] _a = np.exp(__UpperCAmelCase ) / np.exp(__UpperCAmelCase ).sum(-1 , keepdims=__UpperCAmelCase ) _a = list(reversed(scores[0].argsort() ) ) return { "sequence": sequences[0], "labels": [candidate_labels[i] for i in top_inds], "scores": scores[0, top_inds].tolist(), }
362
"""simple docstring""" def A_ ( _lowerCAmelCase : int = 10_00 ): """simple docstring""" return sum(2 * a * ((a - 1) // 2) for a in range(3, n + 1 ) ) if __name__ == "__main__": print(solution())
153
0
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A__ : def __init__( self : Tuple , a : str , a : str=13 , a : Optional[int]=30 , a : Union[str, Any]=2 , a : int=3 , a : Union[str, Any]=True , a : Union[str, Any]=True , a : List[str]=32 , a : int=5 , a : Dict=4 , a : List[str]=37 , a : str="gelu" , a : int=0.1 , a : Union[str, Any]=0.1 , a : Optional[int]=10 , a : Dict=0.0_2 , a : Any=None , a : Union[str, Any]=2 , ): '''simple docstring''' lowerCAmelCase__ : Any = parent lowerCAmelCase__ : List[Any] = batch_size lowerCAmelCase__ : Any = image_size lowerCAmelCase__ : Optional[int] = patch_size lowerCAmelCase__ : Optional[int] = num_channels lowerCAmelCase__ : str = is_training lowerCAmelCase__ : Optional[Any] = use_labels lowerCAmelCase__ : str = hidden_size lowerCAmelCase__ : List[str] = num_hidden_layers lowerCAmelCase__ : List[str] = num_attention_heads lowerCAmelCase__ : Optional[Any] = intermediate_size lowerCAmelCase__ : Dict = hidden_act lowerCAmelCase__ : str = hidden_dropout_prob lowerCAmelCase__ : Optional[Any] = attention_probs_dropout_prob lowerCAmelCase__ : List[str] = type_sequence_label_size lowerCAmelCase__ : Any = initializer_range lowerCAmelCase__ : int = scope lowerCAmelCase__ : Any = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCAmelCase__ : str = (image_size // patch_size) ** 2 lowerCAmelCase__ : Any = num_patches + 1 def _lowerCamelCase ( self : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ : Tuple = None if self.use_labels: lowerCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase__ : List[str] = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self : Dict ): '''simple docstring''' return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=a , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _lowerCamelCase ( self : Optional[int] , a : List[str] , a : List[str] , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Tuple = ViTModel(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : int = model(a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self : Tuple , a : Dict , a : List[str] , a : Tuple ): '''simple docstring''' lowerCAmelCase__ : Optional[int] = ViTForMaskedImageModeling(config=a ) model.to(a ) model.eval() lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ : Optional[Any] = 1 lowerCAmelCase__ : List[str] = ViTForMaskedImageModeling(a ) model.to(a ) model.eval() lowerCAmelCase__ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Union[str, Any] = model(a ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _lowerCamelCase ( self : Any , a : int , a : List[str] , a : Union[str, Any] ): '''simple docstring''' lowerCAmelCase__ : Any = self.type_sequence_label_size lowerCAmelCase__ : Dict = ViTForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : List[str] = model(a , labels=a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCAmelCase__ : List[Any] = 1 lowerCAmelCase__ : Tuple = ViTForImageClassification(a ) model.to(a ) model.eval() lowerCAmelCase__ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ : Any = model(a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ : Tuple = self.prepare_config_and_inputs() ( ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ( lowerCAmelCase__ ) , ) : Any = config_and_inputs lowerCAmelCase__ : Optional[int] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A__ ( __magic_name__ , __magic_name__ , unittest.TestCase ): lowercase = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) lowercase = ( {'feature-extraction': ViTModel, 'image-classification': ViTForImageClassification} if is_torch_available() else {} ) lowercase = True lowercase = False lowercase = False lowercase = False def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : List[Any] = ViTModelTester(self ) lowerCAmelCase__ : str = ConfigTester(self , config_class=a , has_text_modality=a , hidden_size=37 ) def _lowerCamelCase ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='ViT does not use inputs_embeds' ) def _lowerCamelCase ( self : str ): '''simple docstring''' pass def _lowerCamelCase ( self : str ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Union[str, Any] = model_class(a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCAmelCase__ : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a , nn.Linear ) ) def _lowerCamelCase ( self : int ): '''simple docstring''' lowerCAmelCase__ , lowerCAmelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ : Union[str, Any] = model_class(a ) lowerCAmelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ : int = [*signature.parameters.keys()] lowerCAmelCase__ : str = ['pixel_values'] self.assertListEqual(arg_names[:1] , a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a ) def _lowerCamelCase ( self : Dict ): '''simple docstring''' lowerCAmelCase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a ) @slow def _lowerCamelCase ( self : str ): '''simple docstring''' for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ : Optional[Any] = ViTModel.from_pretrained(a ) self.assertIsNotNone(a ) def lowerCAmelCase__ ( ) -> Union[str, Any]: lowerCAmelCase__ : Dict = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self : List[Any] ): '''simple docstring''' return ViTImageProcessor.from_pretrained('google/vit-base-patch16-224' ) if is_vision_available() else None @slow def _lowerCamelCase ( self : Optional[int] ): '''simple docstring''' lowerCAmelCase__ : Tuple = ViTForImageClassification.from_pretrained('google/vit-base-patch16-224' ).to(a ) lowerCAmelCase__ : Optional[int] = self.default_image_processor lowerCAmelCase__ : Union[str, Any] = prepare_img() lowerCAmelCase__ : Optional[Any] = image_processor(images=a , return_tensors='pt' ).to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : Tuple = model(**a ) # verify the logits lowerCAmelCase__ : List[str] = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , a ) lowerCAmelCase__ : Any = torch.tensor([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ).to(a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a , atol=1E-4 ) ) @slow def _lowerCamelCase ( self : List[str] ): '''simple docstring''' lowerCAmelCase__ : Any = ViTModel.from_pretrained('facebook/dino-vits8' ).to(a ) lowerCAmelCase__ : Dict = ViTImageProcessor.from_pretrained('facebook/dino-vits8' , size=480 ) lowerCAmelCase__ : Any = prepare_img() lowerCAmelCase__ : str = image_processor(images=a , return_tensors='pt' ) lowerCAmelCase__ : str = inputs.pixel_values.to(a ) # forward pass with torch.no_grad(): lowerCAmelCase__ : List[str] = model(a , interpolate_pos_encoding=a ) # verify the logits lowerCAmelCase__ : Union[str, Any] = torch.Size((1, 3_601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , a ) lowerCAmelCase__ : Tuple = torch.tensor( [[4.2_3_4_0, 4.3_9_0_6, -6.6_6_9_2], [4.5_4_6_3, 1.8_9_2_8, -6.7_2_5_7], [4.4_4_2_9, 0.8_4_9_6, -5.8_5_8_5]] ).to(a ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , a , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def _lowerCamelCase ( self : Optional[Any] ): '''simple docstring''' lowerCAmelCase__ : int = ViTModel.from_pretrained('facebook/dino-vits8' , torch_dtype=torch.floataa , device_map='auto' ) lowerCAmelCase__ : List[str] = self.default_image_processor lowerCAmelCase__ : Tuple = prepare_img() lowerCAmelCase__ : List[str] = image_processor(images=a , return_tensors='pt' ) lowerCAmelCase__ : Dict = inputs.pixel_values.to(a ) # forward pass to make sure inference works in fp16 with torch.no_grad(): lowerCAmelCase__ : Dict = model(a )
212
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> np.array: lowerCAmelCase__ : Dict = F'''{sampling_rate}''' lowerCAmelCase__ : Any = '1' lowerCAmelCase__ : Optional[Any] = 'f32le' lowerCAmelCase__ : Any = [ 'ffmpeg', '-i', 'pipe:0', '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] try: with subprocess.Popen(SCREAMING_SNAKE_CASE_ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: lowerCAmelCase__ : List[Any] = ffmpeg_process.communicate(SCREAMING_SNAKE_CASE_ ) except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to load audio files from filename' ) from error lowerCAmelCase__ : List[str] = output_stream[0] lowerCAmelCase__ : str = np.frombuffer(SCREAMING_SNAKE_CASE_ , np.floataa ) if audio.shape[0] == 0: raise ValueError('Malformed soundfile' ) return audio def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = "f32le" , ) -> Dict: lowerCAmelCase__ : Optional[Any] = F'''{sampling_rate}''' lowerCAmelCase__ : Any = '1' if format_for_conversion == "s16le": lowerCAmelCase__ : Dict = 2 elif format_for_conversion == "f32le": lowerCAmelCase__ : List[str] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) lowerCAmelCase__ : Tuple = platform.system() if system == "Linux": lowerCAmelCase__ : str = 'alsa' lowerCAmelCase__ : str = 'default' elif system == "Darwin": lowerCAmelCase__ : Any = 'avfoundation' lowerCAmelCase__ : Tuple = ':0' elif system == "Windows": lowerCAmelCase__ : Any = 'dshow' lowerCAmelCase__ : int = 'default' lowerCAmelCase__ : Any = [ 'ffmpeg', '-f', format_, '-i', input_, '-ac', ac, '-ar', ar, '-f', format_for_conversion, '-fflags', 'nobuffer', '-hide_banner', '-loglevel', 'quiet', 'pipe:1', ] lowerCAmelCase__ : List[str] = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample lowerCAmelCase__ : str = _ffmpeg_stream(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for item in iterator: yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = "f32le" , ) -> str: if stream_chunk_s is not None: lowerCAmelCase__ : Union[str, Any] = stream_chunk_s else: lowerCAmelCase__ : Tuple = chunk_length_s lowerCAmelCase__ : Any = ffmpeg_microphone(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , format_for_conversion=SCREAMING_SNAKE_CASE_ ) if format_for_conversion == "s16le": lowerCAmelCase__ : Optional[Any] = np.intaa lowerCAmelCase__ : Optional[Any] = 2 elif format_for_conversion == "f32le": lowerCAmelCase__ : Optional[Any] = np.floataa lowerCAmelCase__ : Optional[Any] = 4 else: raise ValueError(F'''Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`''' ) if stride_length_s is None: lowerCAmelCase__ : Dict = chunk_length_s / 6 lowerCAmelCase__ : int = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(SCREAMING_SNAKE_CASE_ , (int, float) ): lowerCAmelCase__ : Dict = [stride_length_s, stride_length_s] lowerCAmelCase__ : Union[str, Any] = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample lowerCAmelCase__ : List[Any] = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample lowerCAmelCase__ : Any = datetime.datetime.now() lowerCAmelCase__ : Any = datetime.timedelta(seconds=SCREAMING_SNAKE_CASE_ ) for item in chunk_bytes_iter(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , stride=(stride_left, stride_right) , stream=SCREAMING_SNAKE_CASE_ ): # Put everything back in numpy scale lowerCAmelCase__ : Any = np.frombuffer(item['raw'] , dtype=SCREAMING_SNAKE_CASE_ ) lowerCAmelCase__ : Optional[Any] = ( item['stride'][0] // size_of_sample, item['stride'][1] // size_of_sample, ) lowerCAmelCase__ : Optional[int] = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = False ) -> Optional[int]: lowerCAmelCase__ : Union[str, Any] = b'' lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = stride if stride_left + stride_right >= chunk_len: raise ValueError( F'''Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}''' ) lowerCAmelCase__ : List[str] = 0 for raw in iterator: acc += raw if stream and len(SCREAMING_SNAKE_CASE_ ) < chunk_len: lowerCAmelCase__ : Tuple = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(SCREAMING_SNAKE_CASE_ ) >= chunk_len: # We are flushing the accumulator lowerCAmelCase__ : Dict = (_stride_left, stride_right) lowerCAmelCase__ : Any = {'raw': acc[:chunk_len], 'stride': stride} if stream: lowerCAmelCase__ : Optional[int] = False yield item lowerCAmelCase__ : Optional[int] = stride_left lowerCAmelCase__ : Optional[int] = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(SCREAMING_SNAKE_CASE_ ) > stride_left: lowerCAmelCase__ : Tuple = {'raw': acc, 'stride': (_stride_left, 0)} if stream: lowerCAmelCase__ : Any = False yield item def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: lowerCAmelCase__ : int = 2**24 # 16Mo try: with subprocess.Popen(SCREAMING_SNAKE_CASE_ , stdout=subprocess.PIPE , bufsize=SCREAMING_SNAKE_CASE_ ) as ffmpeg_process: while True: lowerCAmelCase__ : List[str] = ffmpeg_process.stdout.read(SCREAMING_SNAKE_CASE_ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('ffmpeg was not found but is required to stream audio files from filename' ) from error
212
1
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', '_float_tensor', 'decoder.output_projection.weight', ] for k in ignore_keys: state_dict.pop(A__, A__ ) def a__ ( A__ ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Any = emb.weight.shape SCREAMING_SNAKE_CASE_ : Optional[int] = nn.Linear(A__, A__, bias=A__ ) SCREAMING_SNAKE_CASE_ : List[str] = emb.weight.data return lin_layer def a__ ( A__, A__="facebook/mbart-large-en-ro", A__=False, A__=False ): SCREAMING_SNAKE_CASE_ : int = torch.load(A__, map_location='cpu' )['model'] remove_ignore_keys_(A__ ) SCREAMING_SNAKE_CASE_ : Optional[Any] = state_dict['encoder.embed_tokens.weight'].shape[0] SCREAMING_SNAKE_CASE_ : Optional[int] = MBartConfig.from_pretrained(A__, vocab_size=A__ ) if mbart_aa and finetuned: SCREAMING_SNAKE_CASE_ : Optional[Any] = 'relu' SCREAMING_SNAKE_CASE_ : Union[str, Any] = state_dict['decoder.embed_tokens.weight'] SCREAMING_SNAKE_CASE_ : int = MBartForConditionalGeneration(A__ ) model.model.load_state_dict(A__ ) if finetuned: SCREAMING_SNAKE_CASE_ : Any = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowerCAmelCase__ : List[str] =argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default='facebook/mbart-large-cc25', type=str, help='Which huggingface architecture to use: mbart-large', ) parser.add_argument('--mbart_50', action='store_true', help='whether the model is mMART-50 checkpoint') parser.add_argument('--finetuned', action='store_true', help='whether the model is a fine-tuned checkpoint') lowerCAmelCase__ : Any =parser.parse_args() lowerCAmelCase__ : Any =convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
162
import numpy # List of input, output pairs lowerCAmelCase__ : int =( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) lowerCAmelCase__ : Any =(((5_15, 22, 13), 5_55), ((61, 35, 49), 1_50)) lowerCAmelCase__ : List[str] =[2, 4, 1, 5] lowerCAmelCase__ : Dict =len(train_data) lowerCAmelCase__ : Union[str, Any] =0.0_0_9 def a__ ( A__, A__="train" ): return calculate_hypothesis_value(A__, A__ ) - output( A__, A__ ) def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : Tuple = 0 for i in range(len(A__ ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def a__ ( A__, A__ ): if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def a__ ( A__, A__ ): if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def a__ ( A__, A__=m ): SCREAMING_SNAKE_CASE_ : Tuple = 0 for i in range(A__ ): if index == -1: summation_value += _error(A__ ) else: summation_value += _error(A__ ) * train_data[i][0][index] return summation_value def a__ ( A__ ): SCREAMING_SNAKE_CASE_ : Any = summation_of_cost_derivative(A__, A__ ) / m return cost_derivative_value def a__ ( ): global parameter_vector # Tune these values to set a tolerance value for predicted output SCREAMING_SNAKE_CASE_ : str = 0.00_00_02 SCREAMING_SNAKE_CASE_ : Any = 0 SCREAMING_SNAKE_CASE_ : Any = 0 while True: j += 1 SCREAMING_SNAKE_CASE_ : int = [0, 0, 0, 0] for i in range(0, len(A__ ) ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = get_cost_derivative(i - 1 ) SCREAMING_SNAKE_CASE_ : str = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( A__, A__, atol=A__, rtol=A__, ): break SCREAMING_SNAKE_CASE_ : Optional[Any] = temp_parameter_vector print(('Number of iterations:', j) ) def a__ ( ): for i in range(len(A__ ) ): print(('Actual output value:', output(A__, 'test' )) ) print(('Hypothesis output:', calculate_hypothesis_value(A__, 'test' )) ) if __name__ == "__main__": run_gradient_descent() print('\nTesting gradient descent for a linear hypothesis function.\n') test_gradient_descent()
162
1
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __snake_case ( __A ,__A ): __lowerCamelCase = 1 @register_to_config def __init__( self , __UpperCamelCase = 1000 , __UpperCamelCase = None ) -> Any: '''simple docstring''' self.set_timesteps(__UpperCamelCase ) # standard deviation of the initial noise distribution snake_case__ : int = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. snake_case__ : str = 4 # running values snake_case__ : Optional[int] = [] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> Optional[int]: '''simple docstring''' snake_case__ : Union[str, Any] = num_inference_steps snake_case__ : Tuple = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] snake_case__ : Dict = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: snake_case__ : Optional[Any] = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: snake_case__ : Tuple = torch.sin(steps * math.pi / 2 ) ** 2 snake_case__ : Any = (1.0 - self.betas**2) ** 0.5 snake_case__ : List[Any] = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] snake_case__ : Optional[Any] = timesteps.to(__UpperCamelCase ) snake_case__ : str = [] def __a ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = True , ) -> Union[SchedulerOutput, Tuple]: '''simple docstring''' if self.num_inference_steps is None: raise ValueError( 'Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler' ) snake_case__ : Dict = (self.timesteps == timestep).nonzero().item() snake_case__ : Tuple = timestep_index + 1 snake_case__ : Union[str, Any] = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__UpperCamelCase ) if len(self.ets ) == 1: snake_case__ : List[Any] = self.ets[-1] elif len(self.ets ) == 2: snake_case__ : Tuple = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: snake_case__ : int = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: snake_case__ : Optional[int] = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) snake_case__ : List[str] = self._get_prev_sample(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__UpperCamelCase ) def __a ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ) -> torch.FloatTensor: '''simple docstring''' return sample def __a ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> Any: '''simple docstring''' snake_case__ : Dict = self.alphas[timestep_index] snake_case__ : Any = self.betas[timestep_index] snake_case__ : str = self.alphas[prev_timestep_index] snake_case__ : str = self.betas[prev_timestep_index] snake_case__ : List[Any] = (sample - sigma * ets) / max(__UpperCamelCase , 1E-8 ) snake_case__ : Optional[int] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self ) -> Union[str, Any]: '''simple docstring''' return self.config.num_train_timesteps
143
def lowerCamelCase__ ( _a , _a): return int((input_a, input_a).count(1) != 0) def lowerCamelCase__ ( ): assert or_gate(0 , 0) == 0 assert or_gate(0 , 1) == 1 assert or_gate(1 , 0) == 1 assert or_gate(1 , 1) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
76
0
from math import factorial def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : int, lowerCAmelCase_ : float ): if successes > trials: raise ValueError('successes must be lower or equal to trials' ) if trials < 0 or successes < 0: raise ValueError('the function is defined for non-negative integers' ) if not isinstance(__a, __a ) or not isinstance(__a, __a ): raise ValueError('the function is defined for non-negative integers' ) if not 0 < prob < 1: raise ValueError('prob has to be in range of 1 - 0' ) __lowerCAmelCase = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! __lowerCAmelCase = float(factorial(__a ) ) coefficient /= factorial(__a ) * factorial(trials - successes ) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print('Probability of 2 successes out of 4 trails') print('with probability of 0.75 is:', end=' ') print(binomial_distribution(2, 4, 0.75))
350
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _snake_case : Union[str, Any] = logging.get_logger(__name__) # pylint: disable=invalid-name def a_ ( lowerCAmelCase_ : Union[List, PIL.Image.Image, torch.Tensor] ): warnings.warn( 'The preprocess method is deprecated and will be removed in a future version. Please' ' use VaeImageProcessor.preprocess instead', lowerCAmelCase_, ) if isinstance(lowerCAmelCase_, torch.Tensor ): return image elif isinstance(lowerCAmelCase_, PIL.Image.Image ): __lowerCAmelCase = [image] if isinstance(image[0], PIL.Image.Image ): __lowerCAmelCase , __lowerCAmelCase = image[0].size __lowerCAmelCase , __lowerCAmelCase = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 __lowerCAmelCase = [np.array(i.resize((w, h), resample=PIL_INTERPOLATION['lanczos'] ) )[None, :] for i in image] __lowerCAmelCase = np.concatenate(lowerCAmelCase_, axis=0 ) __lowerCAmelCase = np.array(lowerCAmelCase_ ).astype(np.floataa ) / 255.0 __lowerCAmelCase = image.transpose(0, 3, 1, 2 ) __lowerCAmelCase = 2.0 * image - 1.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ) elif isinstance(image[0], torch.Tensor ): __lowerCAmelCase = torch.cat(lowerCAmelCase_, dim=0 ) return image def a_ ( lowerCAmelCase_ : Union[List, PIL.Image.Image, torch.Tensor] ): if isinstance(lowerCAmelCase_, torch.Tensor ): return mask elif isinstance(lowerCAmelCase_, PIL.Image.Image ): __lowerCAmelCase = [mask] if isinstance(mask[0], PIL.Image.Image ): __lowerCAmelCase , __lowerCAmelCase = mask[0].size __lowerCAmelCase , __lowerCAmelCase = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 __lowerCAmelCase = [np.array(m.convert('L' ).resize((w, h), resample=PIL_INTERPOLATION['nearest'] ) )[None, :] for m in mask] __lowerCAmelCase = np.concatenate(lowerCAmelCase_, axis=0 ) __lowerCAmelCase = mask.astype(np.floataa ) / 255.0 __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ) elif isinstance(mask[0], torch.Tensor ): __lowerCAmelCase = torch.cat(lowerCAmelCase_, dim=0 ) return mask class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 a_ = 42 def __init__( self : Optional[Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any] ) -> Optional[int]: super().__init__() self.register_modules(unet=lowerCAmelCase_ , scheduler=lowerCAmelCase_ ) @torch.no_grad() def __call__( self : Dict , lowerCAmelCase_ : Union[torch.Tensor, PIL.Image.Image] , lowerCAmelCase_ : Union[torch.Tensor, PIL.Image.Image] , lowerCAmelCase_ : int = 2_5_0 , lowerCAmelCase_ : float = 0.0 , lowerCAmelCase_ : int = 1_0 , lowerCAmelCase_ : int = 1_0 , lowerCAmelCase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCAmelCase_ : Optional[str] = "pil" , lowerCAmelCase_ : bool = True , ) -> Union[ImagePipelineOutput, Tuple]: __lowerCAmelCase = image __lowerCAmelCase = _preprocess_image(lowerCAmelCase_ ) __lowerCAmelCase = original_image.to(device=self.device , dtype=self.unet.dtype ) __lowerCAmelCase = _preprocess_mask(lowerCAmelCase_ ) __lowerCAmelCase = mask_image.to(device=self.device , dtype=self.unet.dtype ) __lowerCAmelCase = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and len(lowerCAmelCase_ ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(lowerCAmelCase_ )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) __lowerCAmelCase = original_image.shape __lowerCAmelCase = randn_tensor(lowerCAmelCase_ , generator=lowerCAmelCase_ , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.device ) __lowerCAmelCase = eta __lowerCAmelCase = self.scheduler.timesteps[0] + 1 __lowerCAmelCase = generator[0] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual __lowerCAmelCase = self.unet(lowerCAmelCase_ , lowerCAmelCase_ ).sample # compute previous image: x_t -> x_t-1 __lowerCAmelCase = self.scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ).prev_sample else: # compute the reverse: x_t-1 -> x_t __lowerCAmelCase = self.scheduler.undo_step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = t __lowerCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __lowerCAmelCase = self.numpy_to_pil(lowerCAmelCase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCAmelCase_ )
207
0
"""simple docstring""" from __future__ import annotations import pandas as pd def _lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_ ): UpperCAmelCase = [0] * no_of_processes UpperCAmelCase = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(lowercase_ ): UpperCAmelCase = burst_time[i] UpperCAmelCase = 0 UpperCAmelCase = 0 UpperCAmelCase = 999999999 UpperCAmelCase = 0 UpperCAmelCase = False # Process until all processes are completed while complete != no_of_processes: for j in range(lowercase_ ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: UpperCAmelCase = remaining_time[j] UpperCAmelCase = j UpperCAmelCase = True if not check: increment_time += 1 continue remaining_time[short] -= 1 UpperCAmelCase = remaining_time[short] if minm == 0: UpperCAmelCase = 999999999 if remaining_time[short] == 0: complete += 1 UpperCAmelCase = False # Find finish time of current process UpperCAmelCase = increment_time + 1 # Calculate waiting time UpperCAmelCase = finish_time - arrival_time[short] UpperCAmelCase = finar - burst_time[short] if waiting_time[short] < 0: UpperCAmelCase = 0 # Increment time increment_time += 1 return waiting_time def _lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_ ): UpperCAmelCase = [0] * no_of_processes for i in range(lowercase_ ): UpperCAmelCase = burst_time[i] + waiting_time[i] return turn_around_time def _lowerCAmelCase ( lowercase_ , lowercase_ , lowercase_ ): UpperCAmelCase = 0 UpperCAmelCase = 0 for i in range(lowercase_ ): UpperCAmelCase = total_waiting_time + waiting_time[i] UpperCAmelCase = total_turn_around_time + turn_around_time[i] print(F"""Average waiting time = {total_waiting_time / no_of_processes:.5f}""" ) print('Average turn around time =' , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("""Enter how many process you want to analyze""") snake_case_ = int(input()) snake_case_ = [0] * no_of_processes snake_case_ = [0] * no_of_processes snake_case_ = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("""Enter the arrival time and burst time for process:--""" + str(i + 1)) snake_case_ , snake_case_ = map(int, input().split()) snake_case_ = calculate_waitingtime(arrival_time, burst_time, no_of_processes) snake_case_ = burst_time snake_case_ = no_of_processes snake_case_ = waiting_time snake_case_ = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) snake_case_ = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ """Process""", """BurstTime""", """ArrivalTime""", """WaitingTime""", """TurnAroundTime""", ], ) # Printing the dataFrame pd.set_option("""display.max_rows""", fcfs.shape[0] + 1) print(fcfs)
78
import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_a, require_torch_gpu from ..pipeline_params import UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS, UNCONDITIONAL_IMAGE_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _lowerCamelCase( _a, unittest.TestCase ): lowercase_ : Union[str, Any] = ConsistencyModelPipeline lowercase_ : Tuple = UNCONDITIONAL_IMAGE_GENERATION_PARAMS lowercase_ : List[str] = UNCONDITIONAL_IMAGE_GENERATION_BATCH_PARAMS # Override required_optional_params to remove num_images_per_prompt lowercase_ : List[str] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """output_type""", """return_dict""", """callback""", """callback_steps""", ] ) @property def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Tuple = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test', subfolder='test_unet', ) return unet @property def UpperCamelCase ( self) -> List[Any]: """simple docstring""" _lowercase : Tuple = UNetaDModel.from_pretrained( 'diffusers/consistency-models-test', subfolder='test_unet_class_cond', ) return unet def UpperCamelCase ( self, lowerCamelCase=False) -> Dict: """simple docstring""" if class_cond: _lowercase : Union[str, Any] = self.dummy_cond_unet else: _lowercase : Union[str, Any] = self.dummy_uncond_unet # Default to CM multistep sampler _lowercase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40, sigma_min=0.0_0_2, sigma_max=8_0.0, ) _lowercase : Optional[Any] = { 'unet': unet, 'scheduler': scheduler, } return components def UpperCamelCase ( self, lowerCamelCase, lowerCamelCase=0) -> Tuple: """simple docstring""" if str(lowerCamelCase).startswith('mps'): _lowercase : str = torch.manual_seed(lowerCamelCase) else: _lowercase : int = torch.Generator(device=lowerCamelCase).manual_seed(lowerCamelCase) _lowercase : Tuple = { 'batch_size': 1, 'num_inference_steps': None, 'timesteps': [22, 0], 'generator': generator, 'output_type': 'np', } return inputs def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Optional[int] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : Optional[int] = self.get_dummy_components() _lowercase : str = ConsistencyModelPipeline(**lowerCamelCase) _lowercase : Dict = pipe.to(lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = self.get_dummy_inputs(lowerCamelCase) _lowercase : Optional[int] = pipe(**lowerCamelCase).images assert image.shape == (1, 32, 32, 3) _lowercase : int = image[0, -3:, -3:, -1] _lowercase : Dict = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> Any: """simple docstring""" _lowercase : Union[str, Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : Dict = self.get_dummy_components(class_cond=lowerCamelCase) _lowercase : Any = ConsistencyModelPipeline(**lowerCamelCase) _lowercase : str = pipe.to(lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = self.get_dummy_inputs(lowerCamelCase) _lowercase : Any = 0 _lowercase : List[str] = pipe(**lowerCamelCase).images assert image.shape == (1, 32, 32, 3) _lowercase : Any = image[0, -3:, -3:, -1] _lowercase : Union[str, Any] = np.array([0.3_5_7_2, 0.6_2_7_3, 0.4_0_3_1, 0.3_9_6_1, 0.4_3_2_1, 0.5_7_3_0, 0.5_2_6_6, 0.4_7_8_0, 0.5_0_0_4]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> Tuple: """simple docstring""" _lowercase : Optional[int] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : Any = self.get_dummy_components() _lowercase : Optional[Any] = ConsistencyModelPipeline(**lowerCamelCase) _lowercase : List[str] = pipe.to(lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Optional[Any] = self.get_dummy_inputs(lowerCamelCase) _lowercase : Union[str, Any] = 1 _lowercase : Tuple = None _lowercase : Tuple = pipe(**lowerCamelCase).images assert image.shape == (1, 32, 32, 3) _lowercase : str = image[0, -3:, -3:, -1] _lowercase : List[str] = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Union[str, Any] = 'cpu' # ensure determinism for the device-dependent torch.Generator _lowercase : Dict = self.get_dummy_components(class_cond=lowerCamelCase) _lowercase : Dict = ConsistencyModelPipeline(**lowerCamelCase) _lowercase : Optional[Any] = pipe.to(lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Tuple = self.get_dummy_inputs(lowerCamelCase) _lowercase : Tuple = 1 _lowercase : int = None _lowercase : Tuple = 0 _lowercase : Dict = pipe(**lowerCamelCase).images assert image.shape == (1, 32, 32, 3) _lowercase : List[str] = image[0, -3:, -3:, -1] _lowercase : Any = np.array([0.5_0_0_4, 0.5_0_0_4, 0.4_9_9_4, 0.5_0_0_8, 0.4_9_7_6, 0.5_0_1_8, 0.4_9_9_0, 0.4_9_8_2, 0.4_9_8_7]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 @slow @require_torch_gpu class _lowerCamelCase( unittest.TestCase ): def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self, lowerCamelCase=0, lowerCamelCase=False, lowerCamelCase="cpu", lowerCamelCase=torch.floataa, lowerCamelCase=(1, 3, 64, 64)) -> Optional[Any]: """simple docstring""" _lowercase : List[Any] = torch.manual_seed(lowerCamelCase) _lowercase : str = { 'num_inference_steps': None, 'timesteps': [22, 0], 'class_labels': 0, 'generator': generator, 'output_type': 'np', } if get_fixed_latents: _lowercase : Optional[Any] = self.get_fixed_latents(seed=lowerCamelCase, device=lowerCamelCase, dtype=lowerCamelCase, shape=lowerCamelCase) _lowercase : Tuple = latents return inputs def UpperCamelCase ( self, lowerCamelCase=0, lowerCamelCase="cpu", lowerCamelCase=torch.floataa, lowerCamelCase=(1, 3, 64, 64)) -> Any: """simple docstring""" if type(lowerCamelCase) == str: _lowercase : Union[str, Any] = torch.device(lowerCamelCase) _lowercase : int = torch.Generator(device=lowerCamelCase).manual_seed(lowerCamelCase) _lowercase : List[str] = randn_tensor(lowerCamelCase, generator=lowerCamelCase, device=lowerCamelCase, dtype=lowerCamelCase) return latents def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : Tuple = UNetaDModel.from_pretrained('diffusers/consistency_models', subfolder='diffusers_cd_imagenet64_l2') _lowercase : Optional[int] = CMStochasticIterativeScheduler( num_train_timesteps=40, sigma_min=0.0_0_2, sigma_max=8_0.0, ) _lowercase : Any = ConsistencyModelPipeline(unet=lowerCamelCase, scheduler=lowerCamelCase) pipe.to(torch_device=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : str = self.get_inputs() _lowercase : Optional[int] = pipe(**lowerCamelCase).images assert image.shape == (1, 64, 64, 3) _lowercase : str = image[0, -3:, -3:, -1] _lowercase : Optional[Any] = np.array([0.0_8_8_8, 0.0_8_8_1, 0.0_6_6_6, 0.0_4_7_9, 0.0_2_9_2, 0.0_1_9_5, 0.0_2_0_1, 0.0_1_6_3, 0.0_2_5_4]) assert np.abs(image_slice.flatten() - expected_slice).max() < 2E-2 def UpperCamelCase ( self) -> str: """simple docstring""" _lowercase : List[str] = UNetaDModel.from_pretrained('diffusers/consistency_models', subfolder='diffusers_cd_imagenet64_l2') _lowercase : List[str] = CMStochasticIterativeScheduler( num_train_timesteps=40, sigma_min=0.0_0_2, sigma_max=8_0.0, ) _lowercase : Union[str, Any] = ConsistencyModelPipeline(unet=lowerCamelCase, scheduler=lowerCamelCase) pipe.to(torch_device=lowerCamelCase) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_inputs() _lowercase : int = 1 _lowercase : Optional[Any] = None _lowercase : str = pipe(**lowerCamelCase).images assert image.shape == (1, 64, 64, 3) _lowercase : List[Any] = image[0, -3:, -3:, -1] _lowercase : List[str] = np.array([0.0_3_4_0, 0.0_1_5_2, 0.0_0_6_3, 0.0_2_6_7, 0.0_2_2_1, 0.0_1_0_7, 0.0_4_1_6, 0.0_1_8_6, 0.0_2_1_7]) assert np.abs(image_slice.flatten() - expected_slice).max() < 2E-2 @require_torch_a def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : str = UNetaDModel.from_pretrained('diffusers/consistency_models', subfolder='diffusers_cd_imagenet64_l2') _lowercase : Optional[int] = CMStochasticIterativeScheduler( num_train_timesteps=40, sigma_min=0.0_0_2, sigma_max=8_0.0, ) _lowercase : Optional[int] = ConsistencyModelPipeline(unet=lowerCamelCase, scheduler=lowerCamelCase) pipe.to(torch_device=lowerCamelCase, torch_dtype=torch.floataa) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : Any = self.get_inputs(get_fixed_latents=lowerCamelCase, device=lowerCamelCase) # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=lowerCamelCase, enable_math=lowerCamelCase, enable_mem_efficient=lowerCamelCase): _lowercase : Dict = pipe(**lowerCamelCase).images assert image.shape == (1, 64, 64, 3) _lowercase : Any = image[0, -3:, -3:, -1] _lowercase : Union[str, Any] = np.array([0.1_8_7_5, 0.1_4_2_8, 0.1_2_8_9, 0.2_1_5_1, 0.2_0_9_2, 0.1_4_7_7, 0.1_8_7_7, 0.1_6_4_1, 0.1_3_5_3]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 @require_torch_a def UpperCamelCase ( self) -> Union[str, Any]: """simple docstring""" _lowercase : Dict = UNetaDModel.from_pretrained('diffusers/consistency_models', subfolder='diffusers_cd_imagenet64_l2') _lowercase : Optional[Any] = CMStochasticIterativeScheduler( num_train_timesteps=40, sigma_min=0.0_0_2, sigma_max=8_0.0, ) _lowercase : int = ConsistencyModelPipeline(unet=lowerCamelCase, scheduler=lowerCamelCase) pipe.to(torch_device=lowerCamelCase, torch_dtype=torch.floataa) pipe.set_progress_bar_config(disable=lowerCamelCase) _lowercase : List[Any] = self.get_inputs(get_fixed_latents=lowerCamelCase, device=lowerCamelCase) _lowercase : int = 1 _lowercase : str = None # Ensure usage of flash attention in torch 2.0 with sdp_kernel(enable_flash=lowerCamelCase, enable_math=lowerCamelCase, enable_mem_efficient=lowerCamelCase): _lowercase : Union[str, Any] = pipe(**lowerCamelCase).images assert image.shape == (1, 64, 64, 3) _lowercase : Any = image[0, -3:, -3:, -1] _lowercase : int = np.array([0.1_6_6_3, 0.1_9_4_8, 0.2_2_7_5, 0.1_6_8_0, 0.1_2_0_4, 0.1_2_4_5, 0.1_8_5_8, 0.1_3_3_8, 0.2_0_9_5]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3
21
0
from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : torch.FloatTensor class a (_lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" @register_to_config def __init__( self : Optional[int] , lowerCamelCase : int = 65536 , lowerCamelCase : Optional[int] = None , lowerCamelCase : int = 2 , lowerCamelCase : int = 2 , lowerCamelCase : int = 0 , lowerCamelCase : str = "fourier" , lowerCamelCase : bool = True , lowerCamelCase : bool = False , lowerCamelCase : float = 0.0 , lowerCamelCase : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , lowerCamelCase : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , lowerCamelCase : Tuple[str] = "UNetMidBlock1D" , lowerCamelCase : str = None , lowerCamelCase : Tuple[int] = (32, 32, 64) , lowerCamelCase : str = None , lowerCamelCase : int = 8 , lowerCamelCase : int = 1 , lowerCamelCase : bool = False , ) -> str: super().__init__() __snake_case : int = sample_size # time if time_embedding_type == "fourier": __snake_case : str = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=lowerCamelCase , log=lowerCamelCase , flip_sin_to_cos=lowerCamelCase ) __snake_case : List[str] = 2 * block_out_channels[0] elif time_embedding_type == "positional": __snake_case : str = Timesteps( block_out_channels[0] , flip_sin_to_cos=lowerCamelCase , downscale_freq_shift=lowerCamelCase ) __snake_case : Union[str, Any] = block_out_channels[0] if use_timestep_embedding: __snake_case : Optional[int] = block_out_channels[0] * 4 __snake_case : Union[str, Any] = TimestepEmbedding( in_channels=lowerCamelCase , time_embed_dim=lowerCamelCase , act_fn=lowerCamelCase , out_dim=block_out_channels[0] , ) __snake_case : List[Any] = nn.ModuleList([] ) __snake_case : Tuple = None __snake_case : Optional[Any] = nn.ModuleList([] ) __snake_case : Union[str, Any] = None # down __snake_case : List[Any] = in_channels for i, down_block_type in enumerate(lowerCamelCase ): __snake_case : str = output_channel __snake_case : Dict = block_out_channels[i] if i == 0: input_channel += extra_in_channels __snake_case : Any = i == len(lowerCamelCase ) - 1 __snake_case : List[str] = get_down_block( lowerCamelCase , num_layers=lowerCamelCase , in_channels=lowerCamelCase , out_channels=lowerCamelCase , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(lowerCamelCase ) # mid __snake_case : str = get_mid_block( lowerCamelCase , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=lowerCamelCase , add_downsample=lowerCamelCase , ) # up __snake_case : List[str] = list(reversed(lowerCamelCase ) ) __snake_case : str = reversed_block_out_channels[0] if out_block_type is None: __snake_case : Tuple = out_channels else: __snake_case : List[Any] = block_out_channels[0] for i, up_block_type in enumerate(lowerCamelCase ): __snake_case : Any = output_channel __snake_case : Dict = ( reversed_block_out_channels[i + 1] if i < len(lowerCamelCase ) - 1 else final_upsample_channels ) __snake_case : List[str] = i == len(lowerCamelCase ) - 1 __snake_case : str = get_up_block( lowerCamelCase , num_layers=lowerCamelCase , in_channels=lowerCamelCase , out_channels=lowerCamelCase , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(lowerCamelCase ) __snake_case : List[str] = output_channel # out __snake_case : Any = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) __snake_case : Any = get_out_block( out_block_type=lowerCamelCase , num_groups_out=lowerCamelCase , embed_dim=block_out_channels[0] , out_channels=lowerCamelCase , act_fn=lowerCamelCase , fc_dim=block_out_channels[-1] // 4 , ) def __snake_case ( self : List[Any] , lowerCamelCase : torch.FloatTensor , lowerCamelCase : Union[torch.Tensor, float, int] , lowerCamelCase : bool = True , ) -> Union[UNetaDOutput, Tuple]: __snake_case : List[Any] = timestep if not torch.is_tensor(lowerCamelCase ): __snake_case : Dict = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(lowerCamelCase ) and len(timesteps.shape ) == 0: __snake_case : List[str] = timesteps[None].to(sample.device ) __snake_case : List[Any] = self.time_proj(lowerCamelCase ) if self.config.use_timestep_embedding: __snake_case : str = self.time_mlp(lowerCamelCase ) else: __snake_case : Any = timestep_embed[..., None] __snake_case : Optional[Any] = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) __snake_case : List[Any] = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down __snake_case : List[str] = () for downsample_block in self.down_blocks: __snake_case , __snake_case : str = downsample_block(hidden_states=lowerCamelCase , temb=lowerCamelCase ) down_block_res_samples += res_samples # 3. mid if self.mid_block: __snake_case : List[str] = self.mid_block(lowerCamelCase , lowerCamelCase ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): __snake_case : Optional[int] = down_block_res_samples[-1:] __snake_case : Optional[int] = down_block_res_samples[:-1] __snake_case : Any = upsample_block(lowerCamelCase , res_hidden_states_tuple=lowerCamelCase , temb=lowerCamelCase ) # 5. post-process if self.out_block: __snake_case : int = self.out_block(lowerCamelCase , lowerCamelCase ) if not return_dict: return (sample,) return UNetaDOutput(sample=lowerCamelCase )
134
import logging from transformers.configuration_utils import PretrainedConfig _snake_case : Optional[Any] = logging.getLogger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Tuple = "masked_bert" def __init__( self : Optional[int] , lowerCamelCase : Any=30522 , lowerCamelCase : Tuple=768 , lowerCamelCase : str=12 , lowerCamelCase : Dict=12 , lowerCamelCase : List[str]=3072 , lowerCamelCase : List[str]="gelu" , lowerCamelCase : Any=0.1 , lowerCamelCase : Any=0.1 , lowerCamelCase : List[Any]=512 , lowerCamelCase : int=2 , lowerCamelCase : str=0.02 , lowerCamelCase : List[str]=1E-12 , lowerCamelCase : Any=0 , lowerCamelCase : Dict="topK" , lowerCamelCase : List[Any]="constant" , lowerCamelCase : Dict=0.0 , **lowerCamelCase : List[Any] , ) -> Dict: super().__init__(pad_token_id=lowerCamelCase , **lowerCamelCase ) __snake_case : Optional[Any] = vocab_size __snake_case : Optional[int] = hidden_size __snake_case : Tuple = num_hidden_layers __snake_case : Union[str, Any] = num_attention_heads __snake_case : Optional[Any] = hidden_act __snake_case : str = intermediate_size __snake_case : Tuple = hidden_dropout_prob __snake_case : Dict = attention_probs_dropout_prob __snake_case : List[str] = max_position_embeddings __snake_case : Union[str, Any] = type_vocab_size __snake_case : Any = initializer_range __snake_case : str = layer_norm_eps __snake_case : str = pruning_method __snake_case : int = mask_init __snake_case : Any = mask_scale
134
1
"""simple docstring""" import numpy as np def snake_case ( A__ ,A__ ,A__ = 1e-12 ,A__ = 1_00 ,): assert np.shape(A__ )[0] == np.shape(A__ )[1] # Ensure proper dimensionality. assert np.shape(A__ )[0] == np.shape(A__ )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(A__ ) == np.iscomplexobj(A__ ) UpperCAmelCase_ : int = np.iscomplexobj(A__ ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(A__ ,input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. UpperCAmelCase_ : Tuple = False UpperCAmelCase_ : str = 0 UpperCAmelCase_ : Optional[int] = 0 UpperCAmelCase_ : Union[str, Any] = 1e12 while not convergence: # Multiple matrix by the vector. UpperCAmelCase_ : Dict = np.dot(A__ ,A__ ) # Normalize the resulting output vector. UpperCAmelCase_ : Optional[int] = w / np.linalg.norm(A__ ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) UpperCAmelCase_ : Optional[int] = vector.conj().T if is_complex else vector.T UpperCAmelCase_ : Optional[int] = np.dot(A__ ,np.dot(A__ ,A__ ) ) # Check convergence. UpperCAmelCase_ : str = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: UpperCAmelCase_ : Union[str, Any] = True UpperCAmelCase_ : int = lambda_ if is_complex: UpperCAmelCase_ : Union[str, Any] = np.real(lambda_ ) return lambda_, vector def snake_case ( ): UpperCAmelCase_ : List[Any] = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) UpperCAmelCase_ : Optional[int] = np.array([41, 4, 20] ) UpperCAmelCase_ : Any = real_input_matrix.astype(np.complexaaa ) UpperCAmelCase_ : List[str] = np.triu(1j * complex_input_matrix ,1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T UpperCAmelCase_ : Any = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": UpperCAmelCase_ : List[str] = real_input_matrix UpperCAmelCase_ : Dict = real_vector elif problem_type == "complex": UpperCAmelCase_ : Dict = complex_input_matrix UpperCAmelCase_ : Dict = complex_vector # Our implementation. UpperCAmelCase_ , UpperCAmelCase_ : Any = power_iteration(A__ ,A__ ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = np.linalg.eigh(A__ ) # Last eigenvalue is the maximum one. UpperCAmelCase_ : List[Any] = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. UpperCAmelCase_ : Optional[int] = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(A__ ) - np.abs(A__ ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
268
"""simple docstring""" # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version lowerCamelCase_ = get_logger(__name__) class UpperCamelCase_ : __magic_name__ = '''dummy_data''' __magic_name__ = '''datasets''' __magic_name__ = False def __init__( self : str , lowerCAmelCase_ : str , lowerCAmelCase_ : str , lowerCAmelCase_ : Union[Version, str] , lowerCAmelCase_ : Optional[str] = None , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : Optional[List[Callable]] = None , ) -> Tuple: UpperCAmelCase_ : Optional[int] = 0 UpperCAmelCase_ : int = dataset_name UpperCAmelCase_ : Optional[int] = cache_dir UpperCAmelCase_ : Tuple = use_local_dummy_data UpperCAmelCase_ : int = config # download_callbacks take a single url as input UpperCAmelCase_ : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root UpperCAmelCase_ : Optional[Any] = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general UpperCAmelCase_ : Dict = str(lowerCAmelCase_ ) # to be downloaded UpperCAmelCase_ : Optional[Any] = None UpperCAmelCase_ : int = None @property def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> str: if self._dummy_file is None: UpperCAmelCase_ : List[str] = self.download_dummy_data() return self._dummy_file @property def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> int: if self.config is not None: # structure is dummy / config_name / version_name return os.path.join("dummy" , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join("dummy" , self.version_name ) @property def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[int]: return os.path.join(self.dummy_data_folder , "dummy_data.zip" ) def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Tuple: UpperCAmelCase_ : int = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) UpperCAmelCase_ : Union[str, Any] = cached_path( lowerCAmelCase_ , cache_dir=self.cache_dir , extract_compressed_file=lowerCAmelCase_ , force_extract=lowerCAmelCase_ ) return os.path.join(lowerCAmelCase_ , self.dummy_file_name ) @property def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> int: return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Optional[Any]: if self._bucket_url is None: UpperCAmelCase_ : Union[str, Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , "/" ) ) return self._bucket_url @property def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[Any]: # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , "/" ).split("/" )[:-1] ) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase_ : List[str] , *lowerCAmelCase_ : List[Any] ) -> Optional[int]: if self.load_existing_dummy_data: # dummy data is downloaded and tested UpperCAmelCase_ : Dict = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned UpperCAmelCase_ : Optional[int] = self.dummy_file_name # special case when data_url is a dict if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): return self.create_dummy_data_dict(lowerCAmelCase_ , lowerCAmelCase_ ) elif isinstance(lowerCAmelCase_ , (list, tuple) ): return self.create_dummy_data_list(lowerCAmelCase_ , lowerCAmelCase_ ) else: return self.create_dummy_data_single(lowerCAmelCase_ , lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase_ : int , *lowerCAmelCase_ : Union[str, Any] ) -> Any: return self.download_and_extract(lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Tuple ) -> Any: return self.download_and_extract(lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase_ : Union[str, Any] , *lowerCAmelCase_ : Tuple , **lowerCAmelCase_ : Tuple ) -> Union[str, Any]: return path def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> List[str]: return {} def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : List[Any] ) -> List[Any]: UpperCAmelCase_ : Dict = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): for single_url in single_urls: download_callback(lowerCAmelCase_ ) else: UpperCAmelCase_ : Tuple = single_urls download_callback(lowerCAmelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase_ : List[str] = [os.path.join(lowerCAmelCase_ , urllib.parse.quote_plus(Path(lowerCAmelCase_ ).name ) ) for x in single_urls] else: UpperCAmelCase_ : Optional[int] = single_urls UpperCAmelCase_ : Optional[Any] = os.path.join(lowerCAmelCase_ , urllib.parse.quote_plus(Path(lowerCAmelCase_ ).name ) ) UpperCAmelCase_ : int = value # make sure that values are unique if all(isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique UpperCAmelCase_ : List[str] = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Optional[int] ) -> Dict: UpperCAmelCase_ : str = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one UpperCAmelCase_ : int = all(bool(re.findall("[0-9]{3,}-of-[0-9]{3,}" , lowerCAmelCase_ ) ) for url in data_url ) UpperCAmelCase_ : Union[str, Any] = all( url.startswith("https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed" ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): UpperCAmelCase_ : Tuple = [data_url[0]] * len(lowerCAmelCase_ ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(lowerCAmelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus UpperCAmelCase_ : Dict = os.path.join(lowerCAmelCase_ , urllib.parse.quote_plus(single_url.split("/" )[-1] ) ) dummy_data_list.append(lowerCAmelCase_ ) return dummy_data_list def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : str ) -> Optional[int]: for download_callback in self.download_callbacks: download_callback(lowerCAmelCase_ ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus UpperCAmelCase_ : Optional[Any] = os.path.join(lowerCAmelCase_ , urllib.parse.quote_plus(data_url.split("/" )[-1] ) ) if os.path.exists(lowerCAmelCase_ ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> int: pass def _SCREAMING_SNAKE_CASE ( self : Any ) -> Optional[int]: pass def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase_ : Dict ) -> Optional[Any]: def _iter_archive_members(lowerCAmelCase_ : Dict ): # this preserves the order of the members inside the ZIP archive UpperCAmelCase_ : str = Path(self.dummy_file ).parent UpperCAmelCase_ : Optional[Any] = path.relative_to(lowerCAmelCase_ ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: UpperCAmelCase_ : str = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(lowerCAmelCase_ ) UpperCAmelCase_ : Optional[Any] = Path(lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = _iter_archive_members(lowerCAmelCase_ ) if self.use_local_dummy_data else path.rglob("*" ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith((".", "__") ): yield file_path.relative_to(lowerCAmelCase_ ).as_posix(), file_path.open("rb" ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase_ : Tuple ) -> str: if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): UpperCAmelCase_ : str = [paths] for path in paths: if os.path.isfile(lowerCAmelCase_ ): if os.path.basename(lowerCAmelCase_ ).startswith((".", "__") ): return yield path else: for dirpath, dirnames, filenames in os.walk(lowerCAmelCase_ ): if os.path.basename(lowerCAmelCase_ ).startswith((".", "__") ): continue dirnames.sort() for filename in sorted(lowerCAmelCase_ ): if filename.startswith((".", "__") ): continue yield os.path.join(lowerCAmelCase_ , lowerCAmelCase_ )
268
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class lowercase_ ( unittest.TestCase ): '''simple docstring''' @slow def __lowerCAmelCase ( self : Optional[int] ) ->Tuple: """simple docstring""" a = AutoModelForSeqaSeqLM.from_pretrained('''google/mt5-small''' , return_dict=__UpperCAmelCase ).to(__UpperCAmelCase ) a = AutoTokenizer.from_pretrained('''google/mt5-small''' ) a = tokenizer('''Hello there''' , return_tensors='''pt''' ).input_ids a = tokenizer('''Hi I am''' , return_tensors='''pt''' ).input_ids a = model(input_ids.to(__UpperCAmelCase ) , labels=labels.to(__UpperCAmelCase ) ).loss a = -(labels.shape[-1] * loss.item()) a = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
26
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowercase_ ( lowercase ): '''simple docstring''' def __init__( self : Union[str, Any] , __UpperCAmelCase : UNetaDModel , __UpperCAmelCase : UNetaDModel , __UpperCAmelCase : DDPMScheduler , __UpperCAmelCase : Optional[int] , ) ->List[str]: """simple docstring""" super().__init__() a = value_function a = unet a = scheduler a = env a = env.get_dataset() a = {} for key in self.data.keys(): try: a = self.data[key].mean() except: # noqa: E722 pass a = {} for key in self.data.keys(): try: a = self.data[key].std() except: # noqa: E722 pass a = env.observation_space.shape[0] a = env.action_space.shape[0] def __lowerCAmelCase ( self : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int] ) ->Dict: """simple docstring""" return (x_in - self.means[key]) / self.stds[key] def __lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : Dict , __UpperCAmelCase : Dict ) ->List[str]: """simple docstring""" return x_in * self.stds[key] + self.means[key] def __lowerCAmelCase ( self : int , __UpperCAmelCase : int ) ->List[str]: """simple docstring""" if type(__UpperCAmelCase ) is dict: return {k: self.to_torch(__UpperCAmelCase ) for k, v in x_in.items()} elif torch.is_tensor(__UpperCAmelCase ): return x_in.to(self.unet.device ) return torch.tensor(__UpperCAmelCase , device=self.unet.device ) def __lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Tuple ) ->int: """simple docstring""" for key, val in cond.items(): a = val.clone() return x_in def __lowerCAmelCase ( self : Optional[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[Any] ) ->Tuple: """simple docstring""" a = x.shape[0] a = None for i in tqdm.tqdm(self.scheduler.timesteps ): # create batch of timesteps to pass into model a = torch.full((batch_size,) , __UpperCAmelCase , device=self.unet.device , dtype=torch.long ) for _ in range(__UpperCAmelCase ): with torch.enable_grad(): x.requires_grad_() # permute to match dimension for pre-trained models a = self.value_function(x.permute(0 , 2 , 1 ) , __UpperCAmelCase ).sample a = torch.autograd.grad([y.sum()] , [x] )[0] a = self.scheduler._get_variance(__UpperCAmelCase ) a = torch.exp(0.5 * posterior_variance ) a = model_std * grad a = 0 a = x.detach() a = x + scale * grad a = self.reset_xa(__UpperCAmelCase , __UpperCAmelCase , self.action_dim ) a = self.unet(x.permute(0 , 2 , 1 ) , __UpperCAmelCase ).sample.permute(0 , 2 , 1 ) # TODO: verify deprecation of this kwarg a = self.scheduler.step(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , predict_epsilon=__UpperCAmelCase )['''prev_sample'''] # apply conditions to the trajectory (set the initial state) a = self.reset_xa(__UpperCAmelCase , __UpperCAmelCase , self.action_dim ) a = self.to_torch(__UpperCAmelCase ) return x, y def __call__( self : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[int]=64 , __UpperCAmelCase : int=32 , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : str=0.1 ) ->List[str]: """simple docstring""" a = self.normalize(__UpperCAmelCase , '''observations''' ) a = obs[None].repeat(__UpperCAmelCase , axis=0 ) a = {0: self.to_torch(__UpperCAmelCase )} a = (batch_size, planning_horizon, self.state_dim + self.action_dim) # generate initial noise and apply our conditions (to make the trajectories start at current state) a = randn_tensor(__UpperCAmelCase , device=self.unet.device ) a = self.reset_xa(__UpperCAmelCase , __UpperCAmelCase , self.action_dim ) a = self.to_torch(__UpperCAmelCase ) # run the diffusion process a , a = self.run_diffusion(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # sort output trajectories by value a = y.argsort(0 , descending=__UpperCAmelCase ).squeeze() a = x[sorted_idx] a = sorted_values[:, :, : self.action_dim] a = actions.detach().cpu().numpy() a = self.de_normalize(__UpperCAmelCase , key='''actions''' ) # select the action with the highest value if y is not None: a = 0 else: # if we didn't run value guiding, select a random action a = np.random.randint(0 , __UpperCAmelCase ) a = denorm_actions[selected_index, 0] return denorm_actions
26
1
def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' while b: UpperCAmelCase__ , UpperCAmelCase__ = b, a % b return a def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' return a if b == 0 else euclidean_gcd_recursive(__A, a % b ) def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3, 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5, 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1, 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3, 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6, 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3, 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5, 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1, 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3, 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6, 3 )}""" ) if __name__ == "__main__": main()
65
'''simple docstring''' import copy import os import cva import numpy as np from matplotlib import pyplot as plt class __UpperCamelCase : def __init__( self ): """simple docstring""" lowerCamelCase_ ='''''' lowerCamelCase_ ='''''' lowerCamelCase_ =[] lowerCamelCase_ =0 lowerCamelCase_ =256 lowerCamelCase_ =0 lowerCamelCase_ =0 lowerCamelCase_ =0 lowerCamelCase_ =0 def lowercase__ ( self, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ =cva.imread(lowerCAmelCase, 0 ) lowerCamelCase_ =copy.deepcopy(self.img ) lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ =plt.hist(self.img.ravel(), 256, [0, 256], label='''x''' ) lowerCamelCase_ =np.sum(lowerCAmelCase ) for i in range(len(lowerCAmelCase ) ): lowerCamelCase_ =x[i] / self.k self.sk += prk lowerCamelCase_ =(self.L - 1) * self.sk if self.rem != 0: lowerCamelCase_ =int(last % last ) lowerCamelCase_ =int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(lowerCAmelCase ) lowerCamelCase_ =int(np.ma.count(self.img ) / self.img[1].size ) lowerCamelCase_ =self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): lowerCamelCase_ =self.img[j][i] if num != self.last_list[num]: lowerCamelCase_ =self.last_list[num] cva.imwrite('''output_data/output.jpg''', self.img ) def lowercase__ ( self ): """simple docstring""" plt.hist(self.img.ravel(), 256, [0, 256] ) def lowercase__ ( self ): """simple docstring""" cva.imshow('''Output-Image''', self.img ) cva.imshow('''Input-Image''', self.original_image ) cva.waitKey(5_000 ) cva.destroyAllWindows() if __name__ == "__main__": a_ : str = os.path.join(os.path.basename(__file__), """image_data/input.jpg""") a_ : Optional[Any] = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
75
0
"""simple docstring""" def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Any: if index == r: for j in range(lowerCAmelCase ): print(data[j] , end=""" """ ) print(""" """ ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location UpperCAmelCase__ : Tuple = arr[i] combination_util(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , index + 1 , lowerCAmelCase , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> Dict: # A temporary array to store all combination one by one UpperCAmelCase__ : Tuple = [0] * r # Print all combination using temporary array 'data[]' combination_util(lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , 0 , lowerCAmelCase , 0 ) if __name__ == "__main__": # Driver code to check the function above _A = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
166
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { """transfo-xl-wt103""": """https://huggingface.co/transfo-xl-wt103/resolve/main/config.json""", } class lowerCamelCase ( lowerCAmelCase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'transfo-xl' SCREAMING_SNAKE_CASE = ['mems'] SCREAMING_SNAKE_CASE = { 'n_token': 'vocab_size', 'hidden_size': 'd_model', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__(self , _lowerCamelCase=267735 , _lowerCamelCase=[20000, 40000, 200000] , _lowerCamelCase=1024 , _lowerCamelCase=1024 , _lowerCamelCase=16 , _lowerCamelCase=64 , _lowerCamelCase=4096 , _lowerCamelCase=4 , _lowerCamelCase=False , _lowerCamelCase=18 , _lowerCamelCase=1600 , _lowerCamelCase=1000 , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase=0 , _lowerCamelCase=-1 , _lowerCamelCase=True , _lowerCamelCase=0.1 , _lowerCamelCase=0.0 , _lowerCamelCase=True , _lowerCamelCase="normal" , _lowerCamelCase=0.01 , _lowerCamelCase=0.01 , _lowerCamelCase=0.02 , _lowerCamelCase=1e-5 , _lowerCamelCase=0 , **_lowerCamelCase , ): """simple docstring""" UpperCAmelCase__ : Any = vocab_size UpperCAmelCase__ : Dict = [] self.cutoffs.extend(_lowerCamelCase ) if proj_share_all_but_first: UpperCAmelCase__ : Optional[int] = [False] + [True] * len(self.cutoffs ) else: UpperCAmelCase__ : List[Any] = [False] + [False] * len(self.cutoffs ) UpperCAmelCase__ : Dict = d_model UpperCAmelCase__ : Dict = d_embed UpperCAmelCase__ : List[Any] = d_head UpperCAmelCase__ : List[str] = d_inner UpperCAmelCase__ : Any = div_val UpperCAmelCase__ : str = pre_lnorm UpperCAmelCase__ : int = n_layer UpperCAmelCase__ : Optional[Any] = n_head UpperCAmelCase__ : Tuple = mem_len UpperCAmelCase__ : Dict = same_length UpperCAmelCase__ : Union[str, Any] = attn_type UpperCAmelCase__ : Optional[int] = clamp_len UpperCAmelCase__ : str = sample_softmax UpperCAmelCase__ : Any = adaptive UpperCAmelCase__ : List[Any] = dropout UpperCAmelCase__ : List[Any] = dropatt UpperCAmelCase__ : Tuple = untie_r UpperCAmelCase__ : str = init UpperCAmelCase__ : Optional[int] = init_range UpperCAmelCase__ : Tuple = proj_init_std UpperCAmelCase__ : str = init_std UpperCAmelCase__ : List[str] = layer_norm_epsilon super().__init__(eos_token_id=_lowerCamelCase , **_lowerCamelCase ) @property def _a (self ): """simple docstring""" logger.info(F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" ) return -1 @max_position_embeddings.setter def _a (self , _lowerCamelCase ): """simple docstring""" raise NotImplementedError( F"""The model {self.model_type} is one of the few models that has no sequence length limit.""" )
166
1
'''simple docstring''' from __future__ import annotations __SCREAMING_SNAKE_CASE : Optional[int] = list[list[int]] # assigning initial values to the grid __SCREAMING_SNAKE_CASE : Matrix = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution __SCREAMING_SNAKE_CASE : Matrix = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def UpperCamelCase_ ( _UpperCAmelCase : Matrix , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> bool: """simple docstring""" for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> tuple[int, int] | None: """simple docstring""" for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> Matrix | None: """simple docstring""" if location := find_empty_location(_UpperCAmelCase ): _UpperCAmelCase , _UpperCAmelCase : int = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 , 10 ): if is_safe(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): _UpperCAmelCase : Dict = digit if sudoku(_UpperCAmelCase ) is not None: return grid _UpperCAmelCase : Tuple = 0 return None def UpperCamelCase_ ( _UpperCAmelCase : Matrix ) -> None: """simple docstring""" for row in grid: for cell in row: print(_UpperCAmelCase , end=" " ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print("""\nExample grid:\n""" + """=""" * 20) print_solution(example_grid) print("""\nExample grid solution:""") __SCREAMING_SNAKE_CASE : Tuple = sudoku(example_grid) if solution is not None: print_solution(solution) else: print("""Cannot find a solution.""")
31
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable __SCREAMING_SNAKE_CASE : Optional[int] = {"""configuration_gpt_neox""": ["""GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXConfig"""]} try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = ["""GPTNeoXTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = [ """GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTNeoXForCausalLM""", """GPTNeoXForQuestionAnswering""", """GPTNeoXForSequenceClassification""", """GPTNeoXForTokenClassification""", """GPTNeoXLayer""", """GPTNeoXModel""", """GPTNeoXPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_neox import GPT_NEOX_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_neox_fast import GPTNeoXTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox import ( GPT_NEOX_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXLayer, GPTNeoXModel, GPTNeoXPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
31
1
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import numpy import tensorflow as tf from transformers.models.esm.modeling_tf_esm import ( TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, TFEsmModel, ) class __lowerCAmelCase : def __init__( self , lowerCAmelCase__ , ) -> int: '''simple docstring''' a__ : Optional[int] =parent a__ : Optional[Any] =1_3 a__ : List[Any] =7 a__ : str =True a__ : Tuple =True a__ : List[Any] =True a__ : Dict =9_9 a__ : Dict =3_2 a__ : Any =2 a__ : Tuple =4 a__ : List[Any] =3_7 a__ : Optional[Any] ='''gelu''' a__ : Tuple =0.1 a__ : str =0.1 a__ : Optional[Any] =5_1_2 a__ : Optional[Any] =1_6 a__ : Optional[int] =2 a__ : int =0.02 a__ : int =3 a__ : Optional[Any] =4 a__ : Optional[Any] =None def _lowercase ( self ) -> int: '''simple docstring''' a__ : Dict =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a__ : Tuple =None if self.use_input_mask: a__ : Union[str, Any] =random_attention_mask([self.batch_size, self.seq_length] ) a__ : Optional[int] =None a__ : Any =None a__ : int =None if self.use_labels: a__ : Tuple =ids_tensor([self.batch_size] , self.type_sequence_label_size ) a__ : Union[str, Any] =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a__ : Optional[Any] =ids_tensor([self.batch_size] , self.num_choices ) a__ : List[str] =EsmConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , pad_token_id=1 , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def _lowercase ( self ) -> List[str]: '''simple docstring''' ( a__ ) : List[Any] =self.prepare_config_and_inputs() a__ : List[str] =True a__ : List[str] =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) a__ : List[Any] =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Any: '''simple docstring''' a__ : List[str] =TFEsmModel(config=_lowerCamelCase ) a__ : int ={'''input_ids''': input_ids, '''attention_mask''': input_mask} a__ : List[str] =model(_lowerCamelCase ) a__ : Any =[input_ids, input_mask] a__ : int =model(_lowerCamelCase ) a__ : List[Any] =model(_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ) -> Optional[int]: '''simple docstring''' a__ : List[str] =True a__ : Tuple =TFEsmModel(config=_lowerCamelCase ) a__ : Tuple ={ '''input_ids''': input_ids, '''attention_mask''': input_mask, '''encoder_hidden_states''': encoder_hidden_states, '''encoder_attention_mask''': encoder_attention_mask, } a__ : str =model(_lowerCamelCase ) a__ : Dict =[input_ids, input_mask] a__ : Union[str, Any] =model(_lowerCamelCase , encoder_hidden_states=_lowerCamelCase ) # Also check the case where encoder outputs are not passed a__ : List[Any] =model(_lowerCamelCase , attention_mask=_lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Union[str, Any]: '''simple docstring''' a__ : Any =TFEsmForMaskedLM(config=_lowerCamelCase ) a__ : int =model([input_ids, input_mask] ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) -> Optional[int]: '''simple docstring''' a__ : List[Any] =self.num_labels a__ : List[Any] =TFEsmForTokenClassification(config=_lowerCamelCase ) a__ : List[str] ={'''input_ids''': input_ids, '''attention_mask''': input_mask} a__ : List[Any] =model(_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _lowercase ( self ) -> Any: '''simple docstring''' a__ : Union[str, Any] =self.prepare_config_and_inputs() ( a__ ) : List[str] =config_and_inputs a__ : Optional[int] ={'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class __lowerCAmelCase ( a__ , a__ , unittest.TestCase): _lowercase : str = ( ( TFEsmModel, TFEsmForMaskedLM, TFEsmForSequenceClassification, TFEsmForTokenClassification, ) if is_tf_available() else () ) _lowercase : Any = ( { 'feature-extraction': TFEsmModel, 'fill-mask': TFEsmForMaskedLM, 'text-classification': TFEsmForSequenceClassification, 'token-classification': TFEsmForTokenClassification, 'zero-shot': TFEsmForSequenceClassification, } if is_tf_available() else {} ) _lowercase : Tuple = False _lowercase : Dict = False def _lowercase ( self ) -> Tuple: '''simple docstring''' a__ : List[Any] =TFEsmModelTester(self ) a__ : Dict =ConfigTester(self , config_class=_lowerCamelCase , hidden_size=3_7 ) def _lowercase ( self ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' a__ : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def _lowercase ( self ) -> int: '''simple docstring''' a__ : Dict =self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*_lowerCamelCase ) def _lowercase ( self ) -> str: '''simple docstring''' a__ : str =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*_lowerCamelCase ) def _lowercase ( self ) -> Optional[Any]: '''simple docstring''' a__ : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*_lowerCamelCase ) @slow def _lowercase ( self ) -> str: '''simple docstring''' for model_name in TF_ESM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : List[str] =TFEsmModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) @unittest.skip("Protein models do not support embedding resizing." ) def _lowercase ( self ) -> Tuple: '''simple docstring''' pass @unittest.skip("Protein models do not support embedding resizing." ) def _lowercase ( self ) -> Union[str, Any]: '''simple docstring''' pass def _lowercase ( self ) -> int: '''simple docstring''' a__ : List[str] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : Tuple =model_class(_lowerCamelCase ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class is TFEsmForMaskedLM: # Output embedding test differs from the main test because they're a matrix, not a layer a__ : Optional[int] =model.get_bias() assert isinstance(_lowerCamelCase , _lowerCamelCase ) for k, v in name.items(): assert isinstance(_lowerCamelCase , tf.Variable ) else: a__ : Any =model.get_output_embeddings() assert x is None a__ : Tuple =model.get_bias() assert name is None @require_tf class __lowerCAmelCase ( unittest.TestCase): @slow def _lowercase ( self ) -> Optional[int]: '''simple docstring''' a__ : Optional[Any] =TFEsmForMaskedLM.from_pretrained("facebook/esm2_t6_8M_UR50D" ) a__ : Union[str, Any] =tf.constant([[0, 1, 2, 3, 4, 5]] ) a__ : Union[str, Any] =model(_lowerCamelCase )[0] a__ : List[Any] =[1, 6, 3_3] self.assertEqual(list(output.numpy().shape ) , _lowerCamelCase ) # compare the actual values for a slice. a__ : Any =tf.constant( [ [ [8.92_15_18, -1_0.5_8_9_8_1_4, -6.4_67_13_07], [-6.3_96_71_56, -1_3.9_1_1_3_7_7, -1.1_21_19_15], [-7.78_12_47, -1_3.9_5_1_5_5_7, -3.74_05_92], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-2 ) ) @slow def _lowercase ( self ) -> Dict: '''simple docstring''' a__ : List[Any] =TFEsmModel.from_pretrained("facebook/esm2_t6_8M_UR50D" ) a__ : Tuple =tf.constant([[0, 6, 4, 1_3, 5, 4, 1_6, 1_2, 1_1, 7, 2]] ) a__ : Any =model(_lowerCamelCase )[0] # compare the actual values for a slice. a__ : int =tf.constant( [ [ [0.14_44_30_92, 0.54_12_53_27, 0.3_24_77_39], [0.30_34_04_84, 0.00_52_66_76, 0.31_07_77_22], [0.32_27_80_43, -0.24_98_70_96, 0.3_41_46_28], ] ] ) self.assertTrue(numpy.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
369
import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class __lowerCAmelCase ( UpperCamelCase__): _lowercase : Any = """char""" _lowercase : str = """bpe""" _lowercase : List[Any] = """wp""" UpperCAmelCase : List[Any] = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class __lowerCAmelCase ( UpperCamelCase__): _lowercase : Tuple = ["""image_processor""", """char_tokenizer"""] _lowercase : Any = """ViTImageProcessor""" _lowercase : Optional[Any] = """MgpstrTokenizer""" def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' a__ : Optional[int] =None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , lowerCAmelCase__ , ) a__ : List[str] =kwargs.pop("feature_extractor" ) a__ : List[str] =image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) a__ : str =tokenizer a__ : List[str] =AutoTokenizer.from_pretrained("gpt2" ) a__ : Optional[int] =AutoTokenizer.from_pretrained("bert-base-uncased" ) super().__init__(lowerCAmelCase__ , lowerCAmelCase__ ) def __call__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=None , **lowerCAmelCase__ ) -> Union[str, Any]: '''simple docstring''' if images is None and text is None: raise ValueError("You need to specify either an `images` or `text` input to process." ) if images is not None: a__ : Union[str, Any] =self.image_processor(lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) if text is not None: a__ : int =self.char_tokenizer(lowerCAmelCase__ , return_tensors=lowerCAmelCase__ , **lowerCAmelCase__ ) if text is None: return inputs elif images is None: return encodings else: a__ : Tuple =encodings["input_ids"] return inputs def _lowercase ( self , lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' a__ , a__ , a__ : Any =sequences a__ : Union[str, Any] =char_preds.size(0 ) a__ , a__ : Dict =self._decode_helper(lowerCAmelCase__ , "char" ) a__ , a__ : List[Any] =self._decode_helper(lowerCAmelCase__ , "bpe" ) a__ , a__ : Optional[int] =self._decode_helper(lowerCAmelCase__ , "wp" ) a__ : List[Any] =[] a__ : Dict =[] for i in range(lowerCAmelCase__ ): a__ : int =[char_scores[i], bpe_scores[i], wp_scores[i]] a__ : Tuple =[char_strs[i], bpe_strs[i], wp_strs[i]] a__ : Any =scores.index(max(lowerCAmelCase__ ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) a__ : Dict ={} a__ : str =final_strs a__ : Optional[int] =final_scores a__ : Union[str, Any] =char_strs a__ : List[str] =bpe_strs a__ : Union[str, Any] =wp_strs return out def _lowercase ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Tuple: '''simple docstring''' if format == DecodeType.CHARACTER: a__ : Optional[Any] =self.char_decode a__ : Dict =1 a__ : Tuple ="[s]" elif format == DecodeType.BPE: a__ : str =self.bpe_decode a__ : Dict =2 a__ : Optional[int] ="#" elif format == DecodeType.WORDPIECE: a__ : Union[str, Any] =self.wp_decode a__ : List[Any] =1_0_2 a__ : Dict ="[SEP]" else: raise ValueError(F'''Format {format} is not supported.''' ) a__ , a__ : Any =[], [] a__ : str =pred_logits.size(0 ) a__ : Optional[Any] =pred_logits.size(1 ) a__ , a__ : Optional[int] =pred_logits.topk(1 , dim=-1 , largest=lowerCAmelCase__ , sorted=lowerCAmelCase__ ) a__ : Optional[Any] =preds_index.view(-1 , lowerCAmelCase__ )[:, 1:] a__ : Dict =decoder(lowerCAmelCase__ ) a__ , a__ : Any =torch.nn.functional.softmax(lowerCAmelCase__ , dim=2 ).max(dim=2 ) a__ : int =preds_max_prob[:, 1:] for index in range(lowerCAmelCase__ ): a__ : Optional[Any] =preds_str[index].find(lowerCAmelCase__ ) a__ : Optional[int] =preds_str[index][:pred_eos] a__ : List[Any] =preds_index[index].cpu().tolist() a__ : List[Any] =pred_index.index(lowerCAmelCase__ ) if eos_token in pred_index else -1 a__ : Union[str, Any] =preds_max_prob[index][: pred_eos_index + 1] a__ : List[Any] =pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(lowerCAmelCase__ ) conf_scores.append(lowerCAmelCase__ ) return dec_strs, conf_scores def _lowercase ( self , lowerCAmelCase__ ) -> str: '''simple docstring''' a__ : int =[seq.replace(" " , "" ) for seq in self.char_tokenizer.batch_decode(lowerCAmelCase__ )] return decode_strs def _lowercase ( self , lowerCAmelCase__ ) -> Tuple: '''simple docstring''' return self.bpe_tokenizer.batch_decode(lowerCAmelCase__ ) def _lowercase ( self , lowerCAmelCase__ ) -> Optional[Any]: '''simple docstring''' a__ : List[Any] =[seq.replace(" " , "" ) for seq in self.wp_tokenizer.batch_decode(lowerCAmelCase__ )] return decode_strs
148
0
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) __UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name __UpperCAmelCase = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior")\n >>> pipe_prior.to("cuda")\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder")\n >>> pipe.to("cuda")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save("cat.png")\n ```\n' def lowercase__ ( __snake_case : int , __snake_case : List[str] , __snake_case : str=8 ): '''simple docstring''' UpperCAmelCase_ : List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase_ : List[Any] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class lowerCamelCase (_snake_case ): '''simple docstring''' def __init__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> Any: super().__init__() self.register_modules( unet=_UpperCamelCase , scheduler=_UpperCamelCase , movq=_UpperCamelCase , ) UpperCAmelCase_ : List[str] = 2 ** (len(self.movq.config.block_out_channels ) - 1) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> int: if latents is None: UpperCAmelCase_ : List[Any] = randn_tensor(_UpperCamelCase , generator=_UpperCamelCase , device=_UpperCamelCase , dtype=_UpperCamelCase ) else: if latents.shape != shape: raise ValueError(f"Unexpected latents shape, got {latents.shape}, expected {shape}" ) UpperCAmelCase_ : Optional[int] = latents.to(_UpperCamelCase ) UpperCAmelCase_ : str = latents * scheduler.init_noise_sigma return latents def __UpperCAmelCase ( self , _UpperCamelCase=0 ) -> int: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) UpperCAmelCase_ : Optional[int] = torch.device(f"cuda:{gpu_id}" ) UpperCAmelCase_ : List[str] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(_UpperCamelCase , _UpperCamelCase ) def __UpperCAmelCase ( self , _UpperCamelCase=0 ) -> List[str]: if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) UpperCAmelCase_ : Any = torch.device(f"cuda:{gpu_id}" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=_UpperCamelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase_ : Dict = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase_ , UpperCAmelCase_ : List[str] = cpu_offload_with_hook(_UpperCamelCase , _UpperCamelCase , prev_module_hook=_UpperCamelCase ) # We'll offload the last model manually. UpperCAmelCase_ : Dict = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __UpperCAmelCase ( self ) -> int: if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(_UpperCamelCase , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(_UpperCamelCase ) def __call__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = 5_1_2 , _UpperCamelCase = 5_1_2 , _UpperCamelCase = 1_0_0 , _UpperCamelCase = 4.0 , _UpperCamelCase = 1 , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = "pil" , _UpperCamelCase = True , ) -> Optional[int]: UpperCAmelCase_ : str = self._execution_device UpperCAmelCase_ : Union[str, Any] = guidance_scale > 1.0 if isinstance(_UpperCamelCase , _UpperCamelCase ): UpperCAmelCase_ : Optional[Any] = torch.cat(_UpperCamelCase , dim=0 ) UpperCAmelCase_ : Optional[int] = image_embeds.shape[0] * num_images_per_prompt if isinstance(_UpperCamelCase , _UpperCamelCase ): UpperCAmelCase_ : Tuple = torch.cat(_UpperCamelCase , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase_ : Optional[Any] = image_embeds.repeat_interleave(_UpperCamelCase , dim=0 ) UpperCAmelCase_ : Dict = negative_image_embeds.repeat_interleave(_UpperCamelCase , dim=0 ) UpperCAmelCase_ : Union[str, Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=_UpperCamelCase ) self.scheduler.set_timesteps(_UpperCamelCase , device=_UpperCamelCase ) UpperCAmelCase_ : List[Any] = self.scheduler.timesteps UpperCAmelCase_ : Any = self.unet.config.in_channels UpperCAmelCase_ , UpperCAmelCase_ : Dict = downscale_height_and_width(_UpperCamelCase , _UpperCamelCase , self.movq_scale_factor ) # create initial latent UpperCAmelCase_ : List[str] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , self.scheduler , ) for i, t in enumerate(self.progress_bar(_UpperCamelCase ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase_ : Optional[int] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase_ : Optional[int] = {'image_embeds': image_embeds} UpperCAmelCase_ : Union[str, Any] = self.unet( sample=_UpperCamelCase , timestep=_UpperCamelCase , encoder_hidden_states=_UpperCamelCase , added_cond_kwargs=_UpperCamelCase , return_dict=_UpperCamelCase , )[0] if do_classifier_free_guidance: UpperCAmelCase_ , UpperCAmelCase_ : List[str] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = noise_pred.chunk(2 ) UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = variance_pred.chunk(2 ) UpperCAmelCase_ : Dict = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase_ : Tuple = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase_ , UpperCAmelCase_ : str = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : Dict = self.scheduler.step( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , generator=_UpperCamelCase , )[0] # post-processing UpperCAmelCase_ : Dict = self.movq.decode(_UpperCamelCase , force_not_quantize=_UpperCamelCase )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}" ) if output_type in ["np", "pil"]: UpperCAmelCase_ : Any = image * 0.5 + 0.5 UpperCAmelCase_ : Optional[Any] = image.clamp(0 , 1 ) UpperCAmelCase_ : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase_ : Tuple = self.numpy_to_pil(_UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_UpperCamelCase )
29
import os from huggingface_hub.constants import HUGGINGFACE_HUB_CACHE, hf_cache_home _UpperCAmelCase : Optional[int] = HUGGINGFACE_HUB_CACHE _UpperCAmelCase : List[str] = "config.json" _UpperCAmelCase : Union[str, Any] = "diffusion_pytorch_model.bin" _UpperCAmelCase : List[Any] = "diffusion_flax_model.msgpack" _UpperCAmelCase : Optional[Any] = "model.onnx" _UpperCAmelCase : int = "diffusion_pytorch_model.safetensors" _UpperCAmelCase : Optional[Any] = "weights.pb" _UpperCAmelCase : Tuple = "https://huggingface.co" _UpperCAmelCase : Union[str, Any] = default_cache_path _UpperCAmelCase : Optional[Any] = "diffusers_modules" _UpperCAmelCase : List[Any] = os.getenv("HF_MODULES_CACHE", os.path.join(hf_cache_home, "modules")) _UpperCAmelCase : Tuple = ["fp16", "non-ema"] _UpperCAmelCase : Any = ".self_attn"
222
0
'''simple docstring''' class lowerCamelCase : '''simple docstring''' def __init__( self : Tuple , lowerCAmelCase_ : int ) -> Any: '''simple docstring''' A__ : Optional[int] =size A__ : Tuple =[0] * size A__ : Union[str, Any] =[0] * size @staticmethod def lowercase__ ( lowerCAmelCase_ : int ) -> Dict: '''simple docstring''' return index | (index + 1) @staticmethod def lowercase__ ( lowerCAmelCase_ : int ) -> str: '''simple docstring''' return (index & (index + 1)) - 1 def lowercase__ ( self : Any , lowerCAmelCase_ : int , lowerCAmelCase_ : int ) -> Tuple: '''simple docstring''' A__ : Any =value while index < self.size: A__ : Optional[Any] =self.get_prev(__lowerCAmelCase ) + 1 if current_left_border == index: A__ : Optional[Any] =value else: A__ : Union[str, Any] =max(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) A__ : Optional[int] =self.get_next(__lowerCAmelCase ) def lowercase__ ( self : Optional[Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : int ) -> Optional[Any]: '''simple docstring''' right -= 1 # Because of right is exclusive A__ : int =0 while left <= right: A__ : Union[str, Any] =self.get_prev(__lowerCAmelCase ) if left <= current_left: A__ : Optional[Any] =max(__lowerCAmelCase , self.tree[right] ) A__ : int =current_left else: A__ : Optional[int] =max(__lowerCAmelCase , self.arr[right] ) right -= 1 return result if __name__ == "__main__": import doctest doctest.testmod()
358
'''simple docstring''' def __lowerCamelCase ( __snake_case : int ) -> bool: """simple docstring""" if p < 2: raise ValueError("""p should not be less than 2!""" ) elif p == 2: return True A__ : Any =4 A__ : int =(1 << p) - 1 for _ in range(p - 2 ): A__ : Dict =((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
136
0
'''simple docstring''' import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class UpperCAmelCase_ (_UpperCAmelCase ): """simple docstring""" lowerCamelCase : BigBirdConfig lowerCamelCase : jnp.dtype = jnp.floataa lowerCamelCase : bool = True def lowercase_ ( self ) -> List[str]: super().setup() __lowerCamelCase : str = nn.Dense(5 , dtype=self.dtype ) def __call__( self , *SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) -> Optional[int]: __lowerCamelCase : Optional[int] = super().__call__(*SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class UpperCAmelCase_ (_UpperCAmelCase ): """simple docstring""" lowerCamelCase : Optional[int] = FlaxBigBirdForNaturalQuestionsModule def UpperCAmelCase__ ( UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] ) -> Tuple: def cross_entropy(UpperCAmelCase_ : int , UpperCAmelCase_ : str , UpperCAmelCase_ : List[str]=None ): __lowerCamelCase : Union[str, Any] = logits.shape[-1] __lowerCamelCase : Any = (labels[..., None] == jnp.arange(UpperCAmelCase_ )[None]).astype('f4' ) __lowerCamelCase : Dict = jax.nn.log_softmax(UpperCAmelCase_ , axis=-1 ) __lowerCamelCase : Optional[int] = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: __lowerCamelCase : str = reduction(UpperCAmelCase_ ) return loss __lowerCamelCase : List[str] = partial(UpperCAmelCase_ , reduction=jnp.mean ) __lowerCamelCase : List[str] = cross_entropy(UpperCAmelCase_ , UpperCAmelCase_ ) __lowerCamelCase : str = cross_entropy(UpperCAmelCase_ , UpperCAmelCase_ ) __lowerCamelCase : Tuple = cross_entropy(UpperCAmelCase_ , UpperCAmelCase_ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class UpperCAmelCase_ : """simple docstring""" lowerCamelCase : str = "google/bigbird-roberta-base" lowerCamelCase : int = 3_0_0_0 lowerCamelCase : int = 1_0_5_0_0 lowerCamelCase : int = 1_2_8 lowerCamelCase : int = 3 lowerCamelCase : int = 1 lowerCamelCase : int = 5 # tx_args lowerCamelCase : float = 3e-5 lowerCamelCase : float = 0.0 lowerCamelCase : int = 2_0_0_0_0 lowerCamelCase : float = 0.0095 lowerCamelCase : str = "bigbird-roberta-natural-questions" lowerCamelCase : str = "training-expt" lowerCamelCase : str = "data/nq-training.jsonl" lowerCamelCase : str = "data/nq-validation.jsonl" def lowercase_ ( self ) -> Dict: os.makedirs(self.base_dir , exist_ok=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = os.path.join(self.base_dir , self.save_dir ) __lowerCamelCase : Dict = self.batch_size_per_device * jax.device_count() @dataclass class UpperCAmelCase_ : """simple docstring""" lowerCamelCase : int lowerCamelCase : int = 4_0_9_6 # no dynamic padding on TPUs def __call__( self , SCREAMING_SNAKE_CASE_ ) -> Dict: __lowerCamelCase : Any = self.collate_fn(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = jax.tree_util.tree_map(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) return batch def lowercase_ ( self , SCREAMING_SNAKE_CASE_ ) -> List[Any]: __lowerCamelCase , __lowerCamelCase : List[str] = self.fetch_inputs(features['input_ids'] ) __lowerCamelCase : Dict = { 'input_ids': jnp.array(SCREAMING_SNAKE_CASE_ , dtype=jnp.intaa ), 'attention_mask': jnp.array(SCREAMING_SNAKE_CASE_ , dtype=jnp.intaa ), 'start_labels': jnp.array(features['start_token'] , dtype=jnp.intaa ), 'end_labels': jnp.array(features['end_token'] , dtype=jnp.intaa ), 'pooled_labels': jnp.array(features['category'] , dtype=jnp.intaa ), } return batch def lowercase_ ( self , SCREAMING_SNAKE_CASE_ ) -> List[str]: __lowerCamelCase : Optional[Any] = [self._fetch_inputs(SCREAMING_SNAKE_CASE_ ) for ids in input_ids] return zip(*SCREAMING_SNAKE_CASE_ ) def lowercase_ ( self , SCREAMING_SNAKE_CASE_ ) -> int: __lowerCamelCase : Tuple = [1 for _ in range(len(SCREAMING_SNAKE_CASE_ ) )] while len(SCREAMING_SNAKE_CASE_ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def UpperCAmelCase__ ( UpperCAmelCase_ : Any , UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[str]=None ) -> Tuple: if seed is not None: __lowerCamelCase : Union[str, Any] = dataset.shuffle(seed=UpperCAmelCase_ ) for i in range(len(UpperCAmelCase_ ) // batch_size ): __lowerCamelCase : List[Any] = dataset[i * batch_size : (i + 1) * batch_size] yield dict(UpperCAmelCase_ ) @partial(jax.pmap , axis_name='batch' ) def UpperCAmelCase__ ( UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[int] , **UpperCAmelCase_ : List[Any] ) -> Dict: def loss_fn(UpperCAmelCase_ : Any ): __lowerCamelCase : Union[str, Any] = model_inputs.pop('start_labels' ) __lowerCamelCase : int = model_inputs.pop('end_labels' ) __lowerCamelCase : int = model_inputs.pop('pooled_labels' ) __lowerCamelCase : int = state.apply_fn(**UpperCAmelCase_ , params=UpperCAmelCase_ , dropout_rng=UpperCAmelCase_ , train=UpperCAmelCase_ ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Optional[Any] = outputs return state.loss_fn( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , ) __lowerCamelCase , __lowerCamelCase : int = jax.random.split(UpperCAmelCase_ ) __lowerCamelCase : List[Any] = jax.value_and_grad(UpperCAmelCase_ ) __lowerCamelCase , __lowerCamelCase : Union[str, Any] = grad_fn(state.params ) __lowerCamelCase : int = jax.lax.pmean({'loss': loss} , axis_name='batch' ) __lowerCamelCase : Dict = jax.lax.pmean(UpperCAmelCase_ , 'batch' ) __lowerCamelCase : Optional[int] = state.apply_gradients(grads=UpperCAmelCase_ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='batch' ) def UpperCAmelCase__ ( UpperCAmelCase_ : List[str] , **UpperCAmelCase_ : int ) -> Union[str, Any]: __lowerCamelCase : Optional[int] = model_inputs.pop('start_labels' ) __lowerCamelCase : List[Any] = model_inputs.pop('end_labels' ) __lowerCamelCase : Optional[Any] = model_inputs.pop('pooled_labels' ) __lowerCamelCase : int = state.apply_fn(**UpperCAmelCase_ , params=state.params , train=UpperCAmelCase_ ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Optional[Any] = outputs __lowerCamelCase : List[str] = state.loss_fn(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) __lowerCamelCase : Optional[int] = jax.lax.pmean({'loss': loss} , axis_name='batch' ) return metrics class UpperCAmelCase_ (train_state.TrainState ): """simple docstring""" lowerCamelCase : Callable = struct.field(pytree_node=_UpperCAmelCase ) @dataclass class UpperCAmelCase_ : """simple docstring""" lowerCamelCase : Args lowerCamelCase : Callable lowerCamelCase : Callable lowerCamelCase : Callable lowerCamelCase : Callable lowerCamelCase : wandb lowerCamelCase : Callable = None def lowercase_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None ) -> Any: __lowerCamelCase : int = model.params __lowerCamelCase : List[Any] = TrainState.create( apply_fn=model.__call__ , params=SCREAMING_SNAKE_CASE_ , tx=SCREAMING_SNAKE_CASE_ , loss_fn=SCREAMING_SNAKE_CASE_ , ) if ckpt_dir is not None: __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Tuple = restore_checkpoint(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = { 'lr': args.lr, 'init_lr': args.init_lr, 'warmup_steps': args.warmup_steps, 'num_train_steps': num_train_steps, 'weight_decay': args.weight_decay, } __lowerCamelCase , __lowerCamelCase : Any = build_tx(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = train_state.TrainState( step=SCREAMING_SNAKE_CASE_ , apply_fn=model.__call__ , params=SCREAMING_SNAKE_CASE_ , tx=SCREAMING_SNAKE_CASE_ , opt_state=SCREAMING_SNAKE_CASE_ , ) __lowerCamelCase : str = args __lowerCamelCase : Tuple = data_collator __lowerCamelCase : int = lr __lowerCamelCase : Optional[Any] = params __lowerCamelCase : Optional[Any] = jax_utils.replicate(SCREAMING_SNAKE_CASE_ ) return state def lowercase_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: __lowerCamelCase : List[Any] = self.args __lowerCamelCase : Optional[Any] = len(SCREAMING_SNAKE_CASE_ ) // args.batch_size __lowerCamelCase : str = jax.random.PRNGKey(0 ) __lowerCamelCase : Optional[Any] = jax.random.split(SCREAMING_SNAKE_CASE_ , jax.device_count() ) for epoch in range(args.max_epochs ): __lowerCamelCase : Union[str, Any] = jnp.array(0 , dtype=jnp.floataa ) __lowerCamelCase : Optional[Any] = get_batched_dataset(SCREAMING_SNAKE_CASE_ , args.batch_size , seed=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : List[Any] = 0 for batch in tqdm(SCREAMING_SNAKE_CASE_ , total=SCREAMING_SNAKE_CASE_ , desc=f'Running EPOCH-{epoch}' ): __lowerCamelCase : List[Any] = self.data_collator(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : str = self.train_step_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) running_loss += jax_utils.unreplicate(metrics['loss'] ) i += 1 if i % args.logging_steps == 0: __lowerCamelCase : Optional[int] = jax_utils.unreplicate(state.step ) __lowerCamelCase : Tuple = running_loss.item() / i __lowerCamelCase : int = self.scheduler_fn(state_step - 1 ) __lowerCamelCase : Tuple = self.evaluate(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Tuple = { 'step': state_step.item(), 'eval_loss': eval_loss.item(), 'tr_loss': tr_loss, 'lr': lr.item(), } tqdm.write(str(SCREAMING_SNAKE_CASE_ ) ) self.logger.log(SCREAMING_SNAKE_CASE_ , commit=SCREAMING_SNAKE_CASE_ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + f'-e{epoch}-s{i}' , state=SCREAMING_SNAKE_CASE_ ) def lowercase_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: __lowerCamelCase : List[Any] = get_batched_dataset(SCREAMING_SNAKE_CASE_ , self.args.batch_size ) __lowerCamelCase : int = len(SCREAMING_SNAKE_CASE_ ) // self.args.batch_size __lowerCamelCase : Any = jnp.array(0 , dtype=jnp.floataa ) __lowerCamelCase : Tuple = 0 for batch in tqdm(SCREAMING_SNAKE_CASE_ , total=SCREAMING_SNAKE_CASE_ , desc='Evaluating ... ' ): __lowerCamelCase : Union[str, Any] = self.data_collator(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = self.val_step_fn(SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) running_loss += jax_utils.unreplicate(metrics['loss'] ) i += 1 return running_loss / i def lowercase_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> int: __lowerCamelCase : Dict = jax_utils.unreplicate(SCREAMING_SNAKE_CASE_ ) print(f'SAVING CHECKPOINT IN {save_dir}' , end=' ... ' ) self.model_save_fn(SCREAMING_SNAKE_CASE_ , params=state.params ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , 'opt_state.msgpack' ) , 'wb' ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args , os.path.join(SCREAMING_SNAKE_CASE_ , 'args.joblib' ) ) joblib.dump(self.data_collator , os.path.join(SCREAMING_SNAKE_CASE_ , 'data_collator.joblib' ) ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , 'training_state.json' ) , 'w' ) as f: json.dump({'step': state.step.item()} , SCREAMING_SNAKE_CASE_ ) print('DONE' ) def UpperCAmelCase__ ( UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Union[str, Any] ) -> Any: print(F'RESTORING CHECKPOINT FROM {save_dir}' , end=' ... ' ) with open(os.path.join(UpperCAmelCase_ , 'flax_model.msgpack' ) , 'rb' ) as f: __lowerCamelCase : Optional[Any] = from_bytes(state.params , f.read() ) with open(os.path.join(UpperCAmelCase_ , 'opt_state.msgpack' ) , 'rb' ) as f: __lowerCamelCase : Optional[int] = from_bytes(state.opt_state , f.read() ) __lowerCamelCase : int = joblib.load(os.path.join(UpperCAmelCase_ , 'args.joblib' ) ) __lowerCamelCase : Optional[Any] = joblib.load(os.path.join(UpperCAmelCase_ , 'data_collator.joblib' ) ) with open(os.path.join(UpperCAmelCase_ , 'training_state.json' ) , 'r' ) as f: __lowerCamelCase : Optional[Any] = json.load(UpperCAmelCase_ ) __lowerCamelCase : Optional[Any] = training_state['step'] print('DONE' ) return params, opt_state, step, args, data_collator def UpperCAmelCase__ ( UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : str , UpperCAmelCase_ : Tuple ) -> Union[str, Any]: __lowerCamelCase : List[Any] = num_train_steps - warmup_steps __lowerCamelCase : Union[str, Any] = optax.linear_schedule(init_value=UpperCAmelCase_ , end_value=UpperCAmelCase_ , transition_steps=UpperCAmelCase_ ) __lowerCamelCase : Dict = optax.linear_schedule(init_value=UpperCAmelCase_ , end_value=1e-7 , transition_steps=UpperCAmelCase_ ) __lowerCamelCase : Optional[int] = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def UpperCAmelCase__ ( UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : int , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[int] ) -> str: def weight_decay_mask(UpperCAmelCase_ : List[Any] ): __lowerCamelCase : List[str] = traverse_util.flatten_dict(UpperCAmelCase_ ) __lowerCamelCase : List[Any] = {k: (v[-1] != 'bias' and v[-2:] != ('LayerNorm', 'scale')) for k, v in params.items()} return traverse_util.unflatten_dict(UpperCAmelCase_ ) __lowerCamelCase : List[Any] = scheduler_fn(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) __lowerCamelCase : Optional[int] = optax.adamw(learning_rate=UpperCAmelCase_ , weight_decay=UpperCAmelCase_ , mask=UpperCAmelCase_ ) return tx, lr
185
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class UpperCAmelCase_ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Union[str, Any] = StableDiffusionInstructPixaPixPipeline lowerCamelCase : List[Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width', 'cross_attention_kwargs'} lowerCamelCase : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS lowerCamelCase : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS lowerCamelCase : List[str] = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase_ ( self ) -> int: torch.manual_seed(0 ) __lowerCamelCase : Any = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=8 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) __lowerCamelCase : Union[str, Any] = PNDMScheduler(skip_prk_steps=SCREAMING_SNAKE_CASE_ ) torch.manual_seed(0 ) __lowerCamelCase : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCamelCase : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) __lowerCamelCase : int = CLIPTextModel(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Optional[int] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) __lowerCamelCase : Optional[Any] = { 'unet': unet, 'scheduler': scheduler, 'vae': vae, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'safety_checker': None, 'feature_extractor': None, } return components def lowercase_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=0 ) -> Dict: __lowerCamelCase : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(SCREAMING_SNAKE_CASE_ ) ).to(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : List[Any] = image.cpu().permute(0 , 2 , 3 , 1 )[0] __lowerCamelCase : int = Image.fromarray(np.uinta(SCREAMING_SNAKE_CASE_ ) ).convert('RGB' ) if str(SCREAMING_SNAKE_CASE_ ).startswith('mps' ): __lowerCamelCase : Union[str, Any] = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) else: __lowerCamelCase : Optional[int] = torch.Generator(device=SCREAMING_SNAKE_CASE_ ).manual_seed(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Tuple = { 'prompt': 'A painting of a squirrel eating a burger', 'image': image, 'generator': generator, 'num_inference_steps': 2, 'guidance_scale': 6.0, 'image_guidance_scale': 1, 'output_type': 'numpy', } return inputs def lowercase_ ( self ) -> Union[str, Any]: __lowerCamelCase : Optional[int] = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Any = self.get_dummy_components() __lowerCamelCase : Tuple = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Optional[int] = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = sd_pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase : Optional[int] = np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> Union[str, Any]: __lowerCamelCase : List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Union[str, Any] = self.get_dummy_components() __lowerCamelCase : List[Any] = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Tuple = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Union[str, Any] = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Union[str, Any] = 'french fries' __lowerCamelCase : List[Any] = sd_pipe(**SCREAMING_SNAKE_CASE_ , negative_prompt=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Tuple = output.images __lowerCamelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __lowerCamelCase : Optional[Any] = np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> Tuple: __lowerCamelCase : List[str] = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : Union[str, Any] = self.get_dummy_components() __lowerCamelCase : Any = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : List[str] = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = [inputs['prompt']] * 2 __lowerCamelCase : Tuple = np.array(inputs['image'] ).astype(np.floataa ) / 2_5_5.0 __lowerCamelCase : List[str] = torch.from_numpy(SCREAMING_SNAKE_CASE_ ).unsqueeze(0 ).to(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = image / 2 + 0.5 __lowerCamelCase : Optional[Any] = image.permute(0 , 3 , 1 , 2 ) __lowerCamelCase : Dict = image.repeat(2 , 1 , 1 , 1 ) __lowerCamelCase : int = sd_pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Tuple = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) __lowerCamelCase : Union[str, Any] = np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> Tuple: __lowerCamelCase : str = 'cpu' # ensure determinism for the device-dependent torch.Generator __lowerCamelCase : int = self.get_dummy_components() __lowerCamelCase : Union[str, Any] = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='scaled_linear' ) __lowerCamelCase : str = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = sd_pipe.to(SCREAMING_SNAKE_CASE_ ) sd_pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = self.get_dummy_inputs(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = sd_pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Optional[Any] = image[0, -3:, -3:, -1] __lowerCamelCase : Tuple = [round(SCREAMING_SNAKE_CASE_ , 4 ) for x in image_slice.flatten().tolist()] print(','.join([str(SCREAMING_SNAKE_CASE_ ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) __lowerCamelCase : List[str] = np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-3 def lowercase_ ( self ) -> List[Any]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def lowercase_ ( self ) -> Optional[int]: __lowerCamelCase : Union[str, Any] = self.get_dummy_components() __lowerCamelCase : Tuple = StableDiffusionInstructPixaPixPipeline(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : int = VaeImageProcessor(do_resize=SCREAMING_SNAKE_CASE_ , do_normalize=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : str = pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : int = pipe(**self.get_dummy_inputs_by_type(SCREAMING_SNAKE_CASE_ , input_image_type='pt' ) )[0] __lowerCamelCase : Optional[Any] = components['vae'] __lowerCamelCase : Dict = self.get_dummy_inputs_by_type(SCREAMING_SNAKE_CASE_ , input_image_type='pt' ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __lowerCamelCase : str = vae.encode(inputs[image_param] ).latent_dist.mode() __lowerCamelCase : str = pipe(**SCREAMING_SNAKE_CASE_ )[0] __lowerCamelCase : Optional[Any] = np.abs(out - out_latents_inputs ).max() self.assertLess(SCREAMING_SNAKE_CASE_ , 1E-4 , 'passing latents as image input generate different result from passing image' ) @slow @require_torch_gpu class UpperCAmelCase_ (unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> int: super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self , SCREAMING_SNAKE_CASE_=0 ) -> str: __lowerCamelCase : str = torch.manual_seed(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = load_image( 'https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg' ) __lowerCamelCase : Any = { 'prompt': 'turn him into a cyborg', 'image': image, 'generator': generator, 'num_inference_steps': 3, 'guidance_scale': 7.5, 'image_guidance_scale': 1.0, 'output_type': 'numpy', } return inputs def lowercase_ ( self ) -> str: __lowerCamelCase : int = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Optional[Any] = self.get_inputs() __lowerCamelCase : List[str] = pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Optional[int] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase : Any = np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowercase_ ( self ) -> Any: __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Optional[Any] = self.get_inputs() __lowerCamelCase : Optional[int] = pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : str = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase : Optional[Any] = np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowercase_ ( self ) -> Optional[int]: __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Dict = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Union[str, Any] = self.get_inputs() __lowerCamelCase : str = pipe(**SCREAMING_SNAKE_CASE_ ).images __lowerCamelCase : Dict = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) __lowerCamelCase : Union[str, Any] = np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1E-3 def lowercase_ ( self ) -> Optional[int]: __lowerCamelCase : Union[str, Any] = 0 def callback_fn(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> None: __lowerCamelCase : List[Any] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __lowerCamelCase : Any = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __lowerCamelCase : Union[str, Any] = latents[0, -3:, -3:, -1] __lowerCamelCase : str = np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 elif step == 2: __lowerCamelCase : Optional[int] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __lowerCamelCase : List[Any] = latents[0, -3:, -3:, -1] __lowerCamelCase : Any = np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5E-2 __lowerCamelCase : int = False __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ , torch_dtype=torch.floataa ) __lowerCamelCase : Optional[Any] = pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Optional[int] = self.get_inputs() pipe(**SCREAMING_SNAKE_CASE_ , callback=SCREAMING_SNAKE_CASE_ , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowercase_ ( self ) -> Union[str, Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __lowerCamelCase : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( 'timbrooks/instruct-pix2pix' , safety_checker=SCREAMING_SNAKE_CASE_ , torch_dtype=torch.floataa ) __lowerCamelCase : List[Any] = pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __lowerCamelCase : List[str] = self.get_inputs() __lowerCamelCase : Tuple = pipe(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Optional[Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def lowercase_ ( self ) -> List[Any]: __lowerCamelCase : Optional[int] = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __lowerCamelCase : Union[str, Any] = inputs['image'].resize((5_04, 5_04) ) __lowerCamelCase : int = 'timbrooks/instruct-pix2pix' __lowerCamelCase : Any = StableDiffusionInstructPixaPixPipeline.from_pretrained( SCREAMING_SNAKE_CASE_ , safety_checker=SCREAMING_SNAKE_CASE_ , ) pipe.to(SCREAMING_SNAKE_CASE_ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE_ ) pipe.enable_attention_slicing() __lowerCamelCase : Dict = pipe(**SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Any = output.images[0] __lowerCamelCase : Optional[int] = image[2_55:2_58, 3_83:3_86, -1] assert image.shape == (5_04, 5_04, 3) __lowerCamelCase : List[str] = np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-3
185
1
def lowerCAmelCase__ ( a__ ) ->str: '''simple docstring''' return "".join(chr(ord(a__ ) - 32 ) if "a" <= char <= "z" else char for char in word ) if __name__ == "__main__": from doctest import testmod testmod()
63
# Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch class _UpperCAmelCase ( TensorFormatter[Mapping, '''torch.Tensor''', Mapping] ): '''simple docstring''' def __init__( self : Union[str, Any] , lowercase_ : Tuple=None , **lowercase_ : Union[str, Any]) -> Union[str, Any]: """simple docstring""" super().__init__(features=lowercase_) _UpperCamelCase = torch_tensor_kwargs import torch # noqa import torch at initialization def __UpperCAmelCase ( self : List[str] , lowercase_ : Union[str, Any]) -> List[str]: """simple docstring""" import torch if isinstance(lowercase_ , lowercase_) and column: if all( isinstance(lowercase_ , torch.Tensor) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column): return torch.stack(lowercase_) return column def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : Union[str, Any]) -> str: """simple docstring""" import torch if isinstance(lowercase_ , (str, bytes, type(lowercase_))): return value elif isinstance(lowercase_ , (np.character, np.ndarray)) and np.issubdtype(value.dtype , np.character): return value.tolist() _UpperCamelCase = {} if isinstance(lowercase_ , (np.number, np.ndarray)) and np.issubdtype(value.dtype , np.integer): _UpperCamelCase = {"dtype": torch.intaa} elif isinstance(lowercase_ , (np.number, np.ndarray)) and np.issubdtype(value.dtype , np.floating): _UpperCamelCase = {"dtype": torch.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(lowercase_ , PIL.Image.Image): _UpperCamelCase = np.asarray(lowercase_) return torch.tensor(lowercase_ , **{**default_dtype, **self.torch_tensor_kwargs}) def __UpperCAmelCase ( self : str , lowercase_ : Dict) -> Dict: """simple docstring""" import torch # support for torch, tf, jax etc. if hasattr(lowercase_ , "__array__") and not isinstance(lowercase_ , torch.Tensor): _UpperCamelCase = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(lowercase_ , np.ndarray): if data_struct.dtype == object: # torch tensors cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(lowercase_) for substruct in data_struct]) elif isinstance(lowercase_ , (list, tuple)): return self._consolidate([self.recursive_tensorize(lowercase_) for substruct in data_struct]) return self._tensorize(lowercase_) def __UpperCAmelCase ( self : Optional[int] , lowercase_ : dict) -> Optional[int]: """simple docstring""" return map_nested(self._recursive_tensorize , lowercase_ , map_list=lowercase_) def __UpperCAmelCase ( self : Optional[Any] , lowercase_ : pa.Table) -> Mapping: """simple docstring""" _UpperCamelCase = self.numpy_arrow_extractor().extract_row(lowercase_) _UpperCamelCase = self.python_features_decoder.decode_row(lowercase_) return self.recursive_tensorize(lowercase_) def __UpperCAmelCase ( self : List[str] , lowercase_ : pa.Table) -> "torch.Tensor": """simple docstring""" _UpperCamelCase = self.numpy_arrow_extractor().extract_column(lowercase_) _UpperCamelCase = self.python_features_decoder.decode_column(lowercase_ , pa_table.column_names[0]) _UpperCamelCase = self.recursive_tensorize(lowercase_) _UpperCamelCase = self._consolidate(lowercase_) return column def __UpperCAmelCase ( self : Tuple , lowercase_ : pa.Table) -> Mapping: """simple docstring""" _UpperCamelCase = self.numpy_arrow_extractor().extract_batch(lowercase_) _UpperCamelCase = self.python_features_decoder.decode_batch(lowercase_) _UpperCamelCase = self.recursive_tensorize(lowercase_) for column_name in batch: _UpperCamelCase = self._consolidate(batch[column_name]) return batch
63
1
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging __lowerCamelCase : str = logging.get_logger(__name__) class A__ ( __snake_case ): _UpperCAmelCase :Optional[Any] = ['pixel_values'] def __init__( self , A_ = True , A_ = 1 / 255 , A_ = True , A_ = 8 , **A_ , ): '''simple docstring''' super().__init__(**A_ ) UpperCamelCase : List[Any] = do_rescale UpperCamelCase : Any = rescale_factor UpperCamelCase : int = do_pad UpperCamelCase : Optional[int] = pad_size def __UpperCamelCase( self , A_ , A_ , A_ = None , **A_ ): '''simple docstring''' return rescale(A_ , scale=A_ , data_format=A_ , **A_ ) def __UpperCamelCase( self , A_ , A_ , A_ = None ): '''simple docstring''' UpperCamelCase , UpperCamelCase : List[str] = get_image_size(A_ ) UpperCamelCase : int = (old_height // size + 1) * size - old_height UpperCamelCase : Dict = (old_width // size + 1) * size - old_width return pad(A_ , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=A_ ) def __UpperCamelCase( self , A_ , A_ = None , A_ = None , A_ = None , A_ = None , A_ = None , A_ = ChannelDimension.FIRST , **A_ , ): '''simple docstring''' UpperCamelCase : List[str] = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase : str = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase : Union[str, Any] = do_pad if do_pad is not None else self.do_pad UpperCamelCase : List[Any] = pad_size if pad_size is not None else self.pad_size UpperCamelCase : Tuple = make_list_of_images(A_ ) if not valid_images(A_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. UpperCamelCase : Any = [to_numpy_array(A_ ) for image in images] if do_rescale: UpperCamelCase : int = [self.rescale(image=A_ , scale=A_ ) for image in images] if do_pad: UpperCamelCase : List[str] = [self.pad(A_ , size=A_ ) for image in images] UpperCamelCase : Dict = [to_channel_dimension_format(A_ , A_ ) for image in images] UpperCamelCase : Dict = {"pixel_values": images} return BatchFeature(data=A_ , tensor_type=A_ )
52
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase : Any = { """configuration_electra""": ["""ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ElectraConfig""", """ElectraOnnxConfig"""], """tokenization_electra""": ["""ElectraTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Dict = ["""ElectraTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Tuple = [ """ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST""", """ElectraForCausalLM""", """ElectraForMaskedLM""", """ElectraForMultipleChoice""", """ElectraForPreTraining""", """ElectraForQuestionAnswering""", """ElectraForSequenceClassification""", """ElectraForTokenClassification""", """ElectraModel""", """ElectraPreTrainedModel""", """load_tf_weights_in_electra""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[Any] = [ """TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFElectraForMaskedLM""", """TFElectraForMultipleChoice""", """TFElectraForPreTraining""", """TFElectraForQuestionAnswering""", """TFElectraForSequenceClassification""", """TFElectraForTokenClassification""", """TFElectraModel""", """TFElectraPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[str] = [ """FlaxElectraForCausalLM""", """FlaxElectraForMaskedLM""", """FlaxElectraForMultipleChoice""", """FlaxElectraForPreTraining""", """FlaxElectraForQuestionAnswering""", """FlaxElectraForSequenceClassification""", """FlaxElectraForTokenClassification""", """FlaxElectraModel""", """FlaxElectraPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys __lowerCamelCase : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
52
1
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class lowercase_ ( pl.LightningModule ): def __init__( self : List[str] , A__ : Union[str, Any] ) -> Any: super().__init__() _snake_case = model _snake_case = 2 _snake_case = nn.Linear(self.model.config.hidden_size , self.num_labels ) def UpperCamelCase_ ( self : Optional[Any] ) -> Optional[Any]: pass def snake_case_(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Dict: """simple docstring""" _snake_case = LongformerModel.from_pretrained(_UpperCamelCase ) _snake_case = LightningModel(_UpperCamelCase ) _snake_case = torch.load(_UpperCamelCase , map_location=torch.device('''cpu''' ) ) lightning_model.load_state_dict(ckpt['''state_dict'''] ) # init longformer question answering model _snake_case = LongformerForQuestionAnswering.from_pretrained(_UpperCamelCase ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(_UpperCamelCase ) print(F"""Conversion successful. Model saved under {pytorch_dump_folder_path}""" ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--longformer_model''', default=None, type=str, required=True, help='''model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.''', ) parser.add_argument( '''--longformer_question_answering_ckpt_path''', default=None, type=str, required=True, help='''Path the official PyTorch Lightning Checkpoint.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __A = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
278
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def snake_case_(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> Optional[Any]: """simple docstring""" _snake_case = AlbertConfig.from_json_file(_UpperCamelCase ) print(F"""Building PyTorch model from configuration: {config}""" ) _snake_case = AlbertForPreTraining(_UpperCamelCase ) # Load weights from tf checkpoint load_tf_weights_in_albert(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) # Save pytorch-model print(F"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _UpperCamelCase ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __A = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
278
1
import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def __magic_name__ ( A : List[Any], A : List[str], A : Optional[Any] ): '''simple docstring''' if isinstance(A, torch.Tensor ): return image elif isinstance(A, PIL.Image.Image ): a = [image] if isinstance(image[0], PIL.Image.Image ): a = [np.array(i.resize((w, h), resample=PIL_INTERPOLATION["lanczos"] ) )[None, :] for i in image] a = np.concatenate(A, axis=0 ) a = np.array(A ).astype(np.floataa ) / 2_55.0 a = image.transpose(0, 3, 1, 2 ) a = 2.0 * image - 1.0 a = torch.from_numpy(A ) elif isinstance(image[0], torch.Tensor ): a = torch.cat(A, dim=0 ) return image def __magic_name__ ( A : str, A : Optional[int], A : Tuple, A : Tuple=0.99_95 ): '''simple docstring''' if not isinstance(A, np.ndarray ): a = True a = va.device a = va.cpu().numpy() a = va.cpu().numpy() a = np.sum(va * va / (np.linalg.norm(A ) * np.linalg.norm(A )) ) if np.abs(A ) > DOT_THRESHOLD: a = (1 - t) * va + t * va else: a = np.arccos(A ) a = np.sin(A ) a = theta_a * t a = np.sin(A ) a = np.sin(theta_a - theta_t ) / sin_theta_a a = sin_theta_t / sin_theta_a a = sa * va + sa * va if inputs_are_torch: a = torch.from_numpy(A ).to(A ) return va def __magic_name__ ( A : int, A : Optional[int] ): '''simple docstring''' a = F.normalize(A, dim=-1 ) a = F.normalize(A, dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def __magic_name__ ( A : List[str], A : str ): '''simple docstring''' for param in model.parameters(): a = value class snake_case__ (_UpperCamelCase ): """simple docstring""" def __init__( self : Optional[Any] , __lowerCamelCase : AutoencoderKL , __lowerCamelCase : CLIPTextModel , __lowerCamelCase : CLIPModel , __lowerCamelCase : CLIPTokenizer , __lowerCamelCase : UNetaDConditionModel , __lowerCamelCase : Union[PNDMScheduler, LMSDiscreteScheduler, DDIMScheduler, DPMSolverMultistepScheduler] , __lowerCamelCase : CLIPFeatureExtractor , __lowerCamelCase : List[Any]=None , __lowerCamelCase : str=None , __lowerCamelCase : Tuple=None , ) -> Optional[int]: super().__init__() self.register_modules( vae=__lowerCamelCase , text_encoder=__lowerCamelCase , clip_model=__lowerCamelCase , tokenizer=__lowerCamelCase , unet=__lowerCamelCase , scheduler=__lowerCamelCase , feature_extractor=__lowerCamelCase , coca_model=__lowerCamelCase , coca_tokenizer=__lowerCamelCase , coca_transform=__lowerCamelCase , ) a = ( feature_extractor.size if isinstance(feature_extractor.size , __lowerCamelCase ) else feature_extractor.size["shortest_edge"] ) a = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , __lowerCamelCase ) set_requires_grad(self.clip_model , __lowerCamelCase ) def __UpperCAmelCase ( self : List[Any] , __lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> List[Any]: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory a = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__lowerCamelCase ) def __UpperCAmelCase ( self : Any ) -> List[Any]: self.enable_attention_slicing(__lowerCamelCase ) def __UpperCAmelCase ( self : List[Any] ) -> Union[str, Any]: set_requires_grad(self.vae , __lowerCamelCase ) def __UpperCAmelCase ( self : int ) -> List[Any]: set_requires_grad(self.vae , __lowerCamelCase ) def __UpperCAmelCase ( self : int ) -> List[Any]: set_requires_grad(self.unet , __lowerCamelCase ) def __UpperCAmelCase ( self : Tuple ) -> str: set_requires_grad(self.unet , __lowerCamelCase ) def __UpperCAmelCase ( self : Optional[Any] , __lowerCamelCase : Dict , __lowerCamelCase : str , __lowerCamelCase : List[str] ) -> Optional[int]: # get the original timestep using init_timestep a = min(int(num_inference_steps * strength ) , __lowerCamelCase ) a = max(num_inference_steps - init_timestep , 0 ) a = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def __UpperCAmelCase ( self : List[str] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Any , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : str=None ) -> Dict: if not isinstance(__lowerCamelCase , torch.Tensor ): raise ValueError(f"""`image` has to be of type `torch.Tensor` but is {type(__lowerCamelCase )}""" ) a = image.to(device=__lowerCamelCase , dtype=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ): a = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(__lowerCamelCase ) ] a = torch.cat(__lowerCamelCase , dim=0 ) else: a = self.vae.encode(__lowerCamelCase ).latent_dist.sample(__lowerCamelCase ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor a = 0.18_215 * init_latents a = init_latents.repeat_interleave(__lowerCamelCase , dim=0 ) a = randn_tensor(init_latents.shape , generator=__lowerCamelCase , device=__lowerCamelCase , dtype=__lowerCamelCase ) # get latents a = self.scheduler.add_noise(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a = init_latents return latents def __UpperCAmelCase ( self : str , __lowerCamelCase : Optional[int] ) -> int: a = self.coca_transform(__lowerCamelCase ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): a = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) a = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split("<end_of_text>" )[0].replace("<start_of_text>" , "" ).rstrip(" .," ) def __UpperCAmelCase ( self : int , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] ) -> Any: a = self.feature_extractor.preprocess(__lowerCamelCase ) a = torch.from_numpy(clip_image_input["pixel_values"][0] ).unsqueeze(0 ).to(self.device ).half() a = self.clip_model.get_image_features(__lowerCamelCase ) a = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__lowerCamelCase ) a = image_embeddings_clip.repeat_interleave(__lowerCamelCase , dim=0 ) return image_embeddings_clip @torch.enable_grad() def __UpperCAmelCase ( self : str , __lowerCamelCase : List[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : str , __lowerCamelCase : Union[str, Any] , ) -> List[str]: a = latents.detach().requires_grad_() a = self.scheduler.scale_model_input(__lowerCamelCase , __lowerCamelCase ) # predict the noise residual a = self.unet(__lowerCamelCase , __lowerCamelCase , encoder_hidden_states=__lowerCamelCase ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): a = self.scheduler.alphas_cumprod[timestep] a = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf a = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 a = torch.sqrt(__lowerCamelCase ) a = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , __lowerCamelCase ): a = self.scheduler.sigmas[index] a = latents - sigma * noise_pred else: raise ValueError(f"""scheduler type {type(self.scheduler )} not supported""" ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor a = 1 / 0.18_215 * sample a = self.vae.decode(__lowerCamelCase ).sample a = (image / 2 + 0.5).clamp(0 , 1 ) a = transforms.Resize(self.feature_extractor_size )(__lowerCamelCase ) a = self.normalize(__lowerCamelCase ).to(latents.dtype ) a = self.clip_model.get_image_features(__lowerCamelCase ) a = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=__lowerCamelCase ) a = spherical_dist_loss(__lowerCamelCase , __lowerCamelCase ).mean() * clip_guidance_scale a = -torch.autograd.grad(__lowerCamelCase , __lowerCamelCase )[0] if isinstance(self.scheduler , __lowerCamelCase ): a = latents.detach() + grads * (sigma**2) a = noise_pred_original else: a = noise_pred_original - torch.sqrt(__lowerCamelCase ) * grads return noise_pred, latents @torch.no_grad() def __call__( self : Optional[int] , __lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , __lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[str] = None , __lowerCamelCase : Optional[int] = 5_12 , __lowerCamelCase : Optional[int] = 5_12 , __lowerCamelCase : float = 0.6 , __lowerCamelCase : Optional[int] = 50 , __lowerCamelCase : Optional[float] = 7.5 , __lowerCamelCase : Optional[int] = 1 , __lowerCamelCase : float = 0.0 , __lowerCamelCase : Optional[float] = 1_00 , __lowerCamelCase : Optional[torch.Generator] = None , __lowerCamelCase : Optional[str] = "pil" , __lowerCamelCase : bool = True , __lowerCamelCase : float = 0.8 , __lowerCamelCase : float = 0.1 , __lowerCamelCase : float = 0.1 , ) -> Dict: if isinstance(__lowerCamelCase , __lowerCamelCase ) and len(__lowerCamelCase ) != batch_size: raise ValueError(f"""You have passed {batch_size} batch_size, but only {len(__lowerCamelCase )} generators.""" ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f"""`height` and `width` have to be divisible by 8 but are {height} and {width}.""" ) if isinstance(__lowerCamelCase , torch.Generator ) and batch_size > 1: a = [generator] + [None] * (batch_size - 1) a = [ ("model", self.coca_model is None), ("tokenizer", self.coca_tokenizer is None), ("transform", self.coca_transform is None), ] a = [x[0] for x in coca_is_none if x[1]] a = ", ".join(__lowerCamelCase ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(__lowerCamelCase ): raise ValueError( f"""Content prompt is None and CoCa [{coca_is_none_str}] is None.""" f"""Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.""" ) a = self.get_image_description(__lowerCamelCase ) if style_prompt is None: if len(__lowerCamelCase ): raise ValueError( f"""Style prompt is None and CoCa [{coca_is_none_str}] is None.""" f""" Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.""" ) a = self.get_image_description(__lowerCamelCase ) # get prompt text embeddings for content and style a = self.tokenizer( __lowerCamelCase , padding="max_length" , max_length=self.tokenizer.model_max_length , truncation=__lowerCamelCase , return_tensors="pt" , ) a = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] a = self.tokenizer( __lowerCamelCase , padding="max_length" , max_length=self.tokenizer.model_max_length , truncation=__lowerCamelCase , return_tensors="pt" , ) a = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] a = slerp(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # duplicate text embeddings for each generation per prompt a = text_embeddings.repeat_interleave(__lowerCamelCase , dim=0 ) # set timesteps a = "offset" in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) a = {} if accepts_offset: a = 1 self.scheduler.set_timesteps(__lowerCamelCase , **__lowerCamelCase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) a , a = self.get_timesteps(__lowerCamelCase , __lowerCamelCase , self.device ) a = timesteps[:1].repeat(__lowerCamelCase ) # Preprocess image a = preprocess(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a = self.prepare_latents( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , text_embeddings.dtype , self.device , __lowerCamelCase ) a = preprocess(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) a = self.prepare_latents( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , text_embeddings.dtype , self.device , __lowerCamelCase ) a = slerp(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if clip_guidance_scale > 0: a = self.get_clip_image_embeddings(__lowerCamelCase , __lowerCamelCase ) a = self.get_clip_image_embeddings(__lowerCamelCase , __lowerCamelCase ) a = slerp( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. a = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: a = content_text_input.input_ids.shape[-1] a = self.tokenizer([""] , padding="max_length" , max_length=__lowerCamelCase , return_tensors="pt" ) a = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt a = uncond_embeddings.repeat_interleave(__lowerCamelCase , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes a = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. a = (batch_size, self.unet.config.in_channels, height // 8, width // 8) a = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps a = torch.randn(__lowerCamelCase , generator=__lowerCamelCase , device="cpu" , dtype=__lowerCamelCase ).to( self.device ) else: a = torch.randn(__lowerCamelCase , generator=__lowerCamelCase , device=self.device , dtype=__lowerCamelCase ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) a = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler a = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] a = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) a = {} if accepts_eta: a = eta # check if the scheduler accepts generator a = "generator" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: a = generator with self.progress_bar(total=__lowerCamelCase ): for i, t in enumerate(__lowerCamelCase ): # expand the latents if we are doing classifier free guidance a = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents a = self.scheduler.scale_model_input(__lowerCamelCase , __lowerCamelCase ) # predict the noise residual a = self.unet(__lowerCamelCase , __lowerCamelCase , encoder_hidden_states=__lowerCamelCase ).sample # perform classifier free guidance if do_classifier_free_guidance: a , a = noise_pred.chunk(2 ) a = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: a = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) a , a = self.cond_fn( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , ) # compute the previous noisy sample x_t -> x_t-1 a = self.scheduler.step(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , **__lowerCamelCase ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor a = 1 / 0.18_215 * latents a = self.vae.decode(__lowerCamelCase ).sample a = (image / 2 + 0.5).clamp(0 , 1 ) a = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": a = self.numpy_to_pil(__lowerCamelCase ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=__lowerCamelCase , nsfw_content_detected=__lowerCamelCase )
107
# XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path A : Optional[Any] = Path(__file__).resolve().parents[3] / 'src' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(4_2) A : List[str] = {'base': 'patrickvonplaten/wav2vec2_tiny_random', 'robust': 'patrickvonplaten/wav2vec2_tiny_random_robust'} A : Optional[int] = 'zero2' A : str = 'zero3' A : Tuple = [ZEROa, ZEROa] def __lowerCAmelCase ( a__ , a__ , a__ ) -> Tuple: # customize the test name generator function as we want both params to appear in the sub-test # name, as by default it shows only the first param __a = parameterized.to_safe_name('''_'''.join(str(a__ ) for x in param.args ) ) return F"""{func.__name__}_{param_based_name}""" # Cartesian-product of zero stages with models to test A : Union[str, Any] = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class __A( a ): @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Any: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) @require_torch_multi_gpu @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> int: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> str: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) @require_torch_multi_gpu @parameterized.expand(_snake_case , name_func=_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Optional[Any]: '''simple docstring''' self.run_and_check( stage=_snake_case , model=_snake_case , distributed=_snake_case , fpaa=_snake_case , ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Tuple: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case = 10 , _snake_case = True , _snake_case = True , _snake_case = True , ) -> Any: '''simple docstring''' __a = models[model] __a = self.run_trainer( stage=_snake_case , model_name=_snake_case , eval_steps=_snake_case , num_train_epochs=1 , distributed=_snake_case , fpaa=_snake_case , ) self.do_checks(_snake_case ) return output_dir def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case , _snake_case = 10 , _snake_case = 1 , _snake_case = True , _snake_case = True , ) -> Union[str, Any]: '''simple docstring''' __a = self.get_auto_remove_tmp_dir('''./xxx''' , after=_snake_case ) __a = F""" --model_name_or_path {model_name} --dataset_name hf-internal-testing/librispeech_asr_dummy --dataset_config_name clean --train_split_name validation --validation_split_name validation --output_dir {output_dir} --num_train_epochs {str(_snake_case )} --per_device_train_batch_size 2 --per_device_eval_batch_size 2 --evaluation_strategy steps --learning_rate 5e-4 --warmup_steps 8 --orthography timit --preprocessing_num_workers 1 --group_by_length --freeze_feature_extractor --report_to none --save_steps 0 --eval_steps {eval_steps} --report_to none """.split() if fpaa: args.extend(['''--fp16'''] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files __a = F"""--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json""".split() __a = [F"""{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"""] __a = self.get_launcher(_snake_case ) __a = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(_snake_case , env=self.get_env() ) return output_dir def SCREAMING_SNAKE_CASE_ ( self , _snake_case=False ) -> List[str]: '''simple docstring''' __a = min(2 , get_gpu_count() ) if distributed else 1 return F"""deepspeed --num_nodes 1 --num_gpus {num_gpus}""".split()
6
0
'''simple docstring''' snake_case__ = { """Pillow""": """Pillow<10.0.0""", """accelerate""": """accelerate>=0.20.3""", """av""": """av==9.2.0""", """beautifulsoup4""": """beautifulsoup4""", """black""": """black~=23.1""", """codecarbon""": """codecarbon==1.2.0""", """cookiecutter""": """cookiecutter==1.7.3""", """dataclasses""": """dataclasses""", """datasets""": """datasets!=2.5.0""", """decord""": """decord==0.6.0""", """deepspeed""": """deepspeed>=0.9.3""", """diffusers""": """diffusers""", """dill""": """dill<0.3.5""", """evaluate""": """evaluate>=0.2.0""", """fairscale""": """fairscale>0.3""", """faiss-cpu""": """faiss-cpu""", """fastapi""": """fastapi""", """filelock""": """filelock""", """flax""": """flax>=0.4.1,<=0.7.0""", """ftfy""": """ftfy""", """fugashi""": """fugashi>=1.0""", """GitPython""": """GitPython<3.1.19""", """hf-doc-builder""": """hf-doc-builder>=0.3.0""", """huggingface-hub""": """huggingface-hub>=0.14.1,<1.0""", """importlib_metadata""": """importlib_metadata""", """ipadic""": """ipadic>=1.0.0,<2.0""", """isort""": """isort>=5.5.4""", """jax""": """jax>=0.2.8,!=0.3.2,<=0.4.13""", """jaxlib""": """jaxlib>=0.1.65,<=0.4.13""", """jieba""": """jieba""", """kenlm""": """kenlm""", """keras-nlp""": """keras-nlp>=0.3.1""", """librosa""": """librosa""", """nltk""": """nltk""", """natten""": """natten>=0.14.6""", """numpy""": """numpy>=1.17""", """onnxconverter-common""": """onnxconverter-common""", """onnxruntime-tools""": """onnxruntime-tools>=1.4.2""", """onnxruntime""": """onnxruntime>=1.4.0""", """opencv-python""": """opencv-python""", """optuna""": """optuna""", """optax""": """optax>=0.0.8,<=0.1.4""", """packaging""": """packaging>=20.0""", """parameterized""": """parameterized""", """phonemizer""": """phonemizer""", """protobuf""": """protobuf""", """psutil""": """psutil""", """pyyaml""": """pyyaml>=5.1""", """pydantic""": """pydantic<2""", """pytest""": """pytest>=7.2.0""", """pytest-timeout""": """pytest-timeout""", """pytest-xdist""": """pytest-xdist""", """python""": """python>=3.8.0""", """ray[tune]""": """ray[tune]""", """regex""": """regex!=2019.12.17""", """requests""": """requests""", """rhoknp""": """rhoknp>=1.1.0,<1.3.1""", """rjieba""": """rjieba""", """rouge-score""": """rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1""", """ruff""": """ruff>=0.0.241,<=0.0.259""", """sacrebleu""": """sacrebleu>=1.4.12,<2.0.0""", """sacremoses""": """sacremoses""", """safetensors""": """safetensors>=0.3.1""", """sagemaker""": """sagemaker>=2.31.0""", """scikit-learn""": """scikit-learn""", """sentencepiece""": """sentencepiece>=0.1.91,!=0.1.92""", """sigopt""": """sigopt""", """starlette""": """starlette""", """sudachipy""": """sudachipy>=0.6.6""", """sudachidict_core""": """sudachidict_core>=20220729""", """tensorflow-cpu""": """tensorflow-cpu>=2.6,<2.14""", """tensorflow""": """tensorflow>=2.6,<2.14""", """tensorflow-text""": """tensorflow-text<2.14""", """tf2onnx""": """tf2onnx""", """timeout-decorator""": """timeout-decorator""", """timm""": """timm""", """tokenizers""": """tokenizers>=0.11.1,!=0.11.3,<0.14""", """torch""": """torch>=1.9,!=1.12.0""", """torchaudio""": """torchaudio""", """torchvision""": """torchvision""", """pyctcdecode""": """pyctcdecode>=0.4.0""", """tqdm""": """tqdm>=4.27""", """unidic""": """unidic>=1.0.2""", """unidic_lite""": """unidic_lite>=1.0.7""", """urllib3""": """urllib3<2.0.0""", """uvicorn""": """uvicorn""", }
356
'''simple docstring''' from __future__ import annotations from PIL import Image # Define glider example snake_case__ = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example snake_case__ = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def snake_case__ ( lowerCamelCase__ : list[list[int]] ) -> list[list[int]]: A_ : str = [] for i in range(len(lowerCamelCase__ ) ): A_ : Optional[Any] = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours A_ : Optional[int] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(lowerCamelCase__ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(lowerCamelCase__ ) - 1: neighbour_count += cells[i + 1][j] if i < len(lowerCamelCase__ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. A_ : List[str] = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(lowerCamelCase__ ) return next_generation def snake_case__ ( lowerCamelCase__ : list[list[int]] , lowerCamelCase__ : int ) -> list[Image.Image]: A_ : List[Any] = [] for _ in range(lowerCamelCase__ ): # Create output image A_ : Optional[int] = Image.new('''RGB''' , (len(cells[0] ), len(lowerCamelCase__ )) ) A_ : int = img.load() # Save cells to image for x in range(len(lowerCamelCase__ ) ): for y in range(len(cells[0] ) ): A_ : Optional[Any] = 2_5_5 - cells[y][x] * 2_5_5 A_ : str = (colour, colour, colour) # Save image images.append(lowerCamelCase__ ) A_ : Optional[int] = new_generation(lowerCamelCase__ ) return images if __name__ == "__main__": snake_case__ = generate_images(GLIDER, 16) images[0].save("""out.gif""", save_all=True, append_images=images[1:])
4
0
"""simple docstring""" import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class __lowerCamelCase ( __lowercase ): '''simple docstring''' def _UpperCAmelCase ( self ) -> Union[str, Any]: _a = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(SCREAMING_SNAKE_CASE__ , '''width_multiplier''' ) ) class __lowerCamelCase : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=64 , __UpperCAmelCase=2 , __UpperCAmelCase=3 , __UpperCAmelCase="swish" , __UpperCAmelCase=3 , __UpperCAmelCase=32 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.02 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=10 , __UpperCAmelCase=None , __UpperCAmelCase=0.25 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , ) -> Optional[Any]: _a = parent _a = batch_size _a = image_size _a = patch_size _a = num_channels _a = make_divisible(512 * width_multiplier , divisor=8 ) _a = hidden_act _a = conv_kernel_size _a = output_stride _a = classifier_dropout_prob _a = use_labels _a = is_training _a = num_labels _a = initializer_range _a = scope _a = width_multiplier _a = ffn_dropout _a = attn_dropout def _UpperCAmelCase ( self ) -> Optional[Any]: _a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _a = None _a = None if self.use_labels: _a = ids_tensor([self.batch_size] , self.num_labels ) _a = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _a = self.get_config() return config, pixel_values, labels, pixel_labels def _UpperCAmelCase ( self ) -> int: return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Any: _a = MobileViTVaModel(config=SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() _a = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: _a = self.num_labels _a = MobileViTVaForImageClassification(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() _a = model(SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> Any: _a = self.num_labels _a = MobileViTVaForSemanticSegmentation(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() _a = model(SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) _a = model(SCREAMING_SNAKE_CASE__ , labels=SCREAMING_SNAKE_CASE__ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def _UpperCAmelCase ( self ) -> Dict: _a = self.prepare_config_and_inputs() _a , _a , _a , _a = config_and_inputs _a = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __lowerCamelCase ( __lowercase , __lowercase , unittest.TestCase ): '''simple docstring''' A_ : Optional[int] = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) A_ : Any = ( { """feature-extraction""": MobileViTVaModel, """image-classification""": MobileViTVaForImageClassification, """image-segmentation""": MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) A_ : List[str] = False A_ : Optional[Any] = False A_ : str = False A_ : Optional[Any] = False def _UpperCAmelCase ( self ) -> Optional[Any]: _a = MobileViTVaModelTester(self ) _a = MobileViTVaConfigTester(self , config_class=SCREAMING_SNAKE_CASE__ , has_text_modality=SCREAMING_SNAKE_CASE__ ) def _UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() @unittest.skip(reason='''MobileViTV2 does not use inputs_embeds''' ) def _UpperCAmelCase ( self ) -> str: pass @unittest.skip(reason='''MobileViTV2 does not support input and output embeddings''' ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass @unittest.skip(reason='''MobileViTV2 does not output attentions''' ) def _UpperCAmelCase ( self ) -> Tuple: pass @require_torch_multi_gpu @unittest.skip(reason='''Got `CUDA error: misaligned address` for tests after this one being run.''' ) def _UpperCAmelCase ( self ) -> List[Any]: pass @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def _UpperCAmelCase ( self ) -> Dict: pass def _UpperCAmelCase ( self ) -> Any: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = model_class(SCREAMING_SNAKE_CASE__ ) _a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _a = [*signature.parameters.keys()] _a = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE__ ) def _UpperCAmelCase ( self ) -> Any: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE__ ) def _UpperCAmelCase ( self ) -> Optional[Any]: def check_hidden_states_output(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): _a = model_class(SCREAMING_SNAKE_CASE__ ) model.to(SCREAMING_SNAKE_CASE__ ) model.eval() with torch.no_grad(): _a = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) _a = outputs.hidden_states _a = 5 self.assertEqual(len(SCREAMING_SNAKE_CASE__ ) , SCREAMING_SNAKE_CASE__ ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. _a = 2 for i in range(len(SCREAMING_SNAKE_CASE__ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) _a , _a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _a = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _a = True check_hidden_states_output(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) def _UpperCAmelCase ( self ) -> Optional[Any]: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*SCREAMING_SNAKE_CASE__ ) def _UpperCAmelCase ( self ) -> Tuple: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*SCREAMING_SNAKE_CASE__ ) @slow def _UpperCAmelCase ( self ) -> Any: for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _a = MobileViTVaModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE__ ) def A_ ( ): """simple docstring""" _a = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def _UpperCAmelCase ( self ) -> Dict: return ( MobileViTImageProcessor.from_pretrained('''apple/mobilevitv2-1.0-imagenet1k-256''' ) if is_vision_available() else None ) @slow def _UpperCAmelCase ( self ) -> Union[str, Any]: _a = MobileViTVaForImageClassification.from_pretrained('''apple/mobilevitv2-1.0-imagenet1k-256''' ).to( SCREAMING_SNAKE_CASE__ ) _a = self.default_image_processor _a = prepare_img() _a = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): _a = model(**SCREAMING_SNAKE_CASE__ ) # verify the logits _a = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE__ ) _a = torch.tensor([-1.6_3_3_6e0_0, -7.3_2_0_4e-0_2, -5.1_8_8_3e-0_1] ).to(SCREAMING_SNAKE_CASE__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) ) @slow def _UpperCAmelCase ( self ) -> Any: _a = MobileViTVaForSemanticSegmentation.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) _a = model.to(SCREAMING_SNAKE_CASE__ ) _a = MobileViTImageProcessor.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) _a = prepare_img() _a = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): _a = model(**SCREAMING_SNAKE_CASE__ ) _a = outputs.logits # verify the logits _a = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , SCREAMING_SNAKE_CASE__ ) _a = torch.tensor( [ [[7.0863, 7.1525, 6.8201], [6.6931, 6.8770, 6.8933], [6.2978, 7.0366, 6.9636]], [[-3.7134, -3.6712, -3.6675], [-3.5825, -3.3549, -3.4777], [-3.3435, -3.3979, -3.2857]], [[-2.9329, -2.8003, -2.7369], [-3.0564, -2.4780, -2.0207], [-2.6889, -1.9298, -1.7640]], ] , device=SCREAMING_SNAKE_CASE__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , SCREAMING_SNAKE_CASE__ , atol=1e-4 ) ) @slow def _UpperCAmelCase ( self ) -> Optional[int]: _a = MobileViTVaForSemanticSegmentation.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) _a = model.to(SCREAMING_SNAKE_CASE__ ) _a = MobileViTImageProcessor.from_pretrained('''shehan97/mobilevitv2-1.0-voc-deeplabv3''' ) _a = prepare_img() _a = image_processor(images=SCREAMING_SNAKE_CASE__ , return_tensors='''pt''' ).to(SCREAMING_SNAKE_CASE__ ) # forward pass with torch.no_grad(): _a = model(**SCREAMING_SNAKE_CASE__ ) _a = outputs.logits.detach().cpu() _a = image_processor.post_process_semantic_segmentation(outputs=SCREAMING_SNAKE_CASE__ , target_sizes=[(50, 60)] ) _a = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , SCREAMING_SNAKE_CASE__ ) _a = image_processor.post_process_semantic_segmentation(outputs=SCREAMING_SNAKE_CASE__ ) _a = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , SCREAMING_SNAKE_CASE__ )
320
from typing import Optional from .. import Features, NamedSplit from ..packaged_modules.text.text import Text from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowerCAmelCase__ ( __lowercase ): def __init__( self : Tuple , SCREAMING_SNAKE_CASE__ : NestedDataStructureLike[PathLike] , SCREAMING_SNAKE_CASE__ : Optional[NamedSplit] = None , SCREAMING_SNAKE_CASE__ : Optional[Features] = None , SCREAMING_SNAKE_CASE__ : str = None , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : bool = False , SCREAMING_SNAKE_CASE__ : Optional[int] = None , **SCREAMING_SNAKE_CASE__ : str , ) -> Union[str, Any]: super().__init__( SCREAMING_SNAKE_CASE__ , split=SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , cache_dir=SCREAMING_SNAKE_CASE__ , keep_in_memory=SCREAMING_SNAKE_CASE__ , streaming=SCREAMING_SNAKE_CASE__ , num_proc=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) __lowerCamelCase = path_or_paths if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else {self.split: path_or_paths} __lowerCamelCase = Text( cache_dir=SCREAMING_SNAKE_CASE__ , data_files=SCREAMING_SNAKE_CASE__ , features=SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ , ) def __A ( self : int ) -> Dict: # Build iterable dataset if self.streaming: __lowerCamelCase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None __lowerCamelCase = None self.builder.download_and_prepare( download_config=SCREAMING_SNAKE_CASE__ , download_mode=SCREAMING_SNAKE_CASE__ , verification_mode=SCREAMING_SNAKE_CASE__ , base_path=SCREAMING_SNAKE_CASE__ , num_proc=self.num_proc , ) __lowerCamelCase = self.builder.as_dataset( split=self.split , verification_mode=SCREAMING_SNAKE_CASE__ , in_memory=self.keep_in_memory ) return dataset
270
0
"""simple docstring""" from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time __A = Lock() def __A (_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->Optional[int]: """simple docstring""" global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(_SCREAMING_SNAKE_CASE ) process_lock.release() # receive your right neighbor's value process_lock.acquire() lowerCAmelCase__ :Optional[int] = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left lowerCAmelCase__ :str = min(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(_SCREAMING_SNAKE_CASE ) process_lock.release() # receive your left neighbor's value process_lock.acquire() lowerCAmelCase__ :Any = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right lowerCAmelCase__ :Optional[Any] = max(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # after all swaps are performed, send the values back to main result_pipe[1].send(_SCREAMING_SNAKE_CASE ) def __A (_SCREAMING_SNAKE_CASE ) ->Dict: """simple docstring""" lowerCAmelCase__ :Tuple = [] lowerCAmelCase__ :Dict = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop lowerCAmelCase__ :int = Pipe() lowerCAmelCase__ :Optional[int] = Pipe() process_array_.append( Process( target=_SCREAMING_SNAKE_CASE , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) lowerCAmelCase__ :Optional[Any] = temp_rs lowerCAmelCase__ :str = temp_rr for i in range(1 , len(_SCREAMING_SNAKE_CASE ) - 1 ): lowerCAmelCase__ :Tuple = Pipe() lowerCAmelCase__ :Tuple = Pipe() process_array_.append( Process( target=_SCREAMING_SNAKE_CASE , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) lowerCAmelCase__ :Optional[int] = temp_rs lowerCAmelCase__ :List[str] = temp_rr process_array_.append( Process( target=_SCREAMING_SNAKE_CASE , args=( len(_SCREAMING_SNAKE_CASE ) - 1, arr[len(_SCREAMING_SNAKE_CASE ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(_SCREAMING_SNAKE_CASE ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(_SCREAMING_SNAKE_CASE ) ): lowerCAmelCase__ :Any = result_pipe[p][0].recv() process_array_[p].join() return arr def __A () ->Union[str, Any]: """simple docstring""" lowerCAmelCase__ :Optional[Any] = list(range(10 , 0 , -1 ) ) print('Initial List' ) print(*_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ :str = odd_even_transposition(_SCREAMING_SNAKE_CASE ) print('Sorted List\n' ) print(*_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
254
"""simple docstring""" from __future__ import annotations import csv import requests from bsa import BeautifulSoup def __A (_SCREAMING_SNAKE_CASE = "" ) ->dict[str, float]: """simple docstring""" lowerCAmelCase__ :Optional[Any] = url or 'https://www.imdb.com/chart/top/?ref_=nv_mv_250' lowerCAmelCase__ :str = BeautifulSoup(requests.get(_SCREAMING_SNAKE_CASE ).text , 'html.parser' ) lowerCAmelCase__ :List[Any] = soup.find_all('td' , attrs='titleColumn' ) lowerCAmelCase__ :Optional[int] = soup.find_all('td' , class_='ratingColumn imdbRating' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) } def __A (_SCREAMING_SNAKE_CASE = "IMDb_Top_250_Movies.csv" ) ->None: """simple docstring""" lowerCAmelCase__ :Any = get_imdb_top_aaa_movies() with open(_SCREAMING_SNAKE_CASE , 'w' , newline='' ) as out_file: lowerCAmelCase__ :Dict = csv.writer(_SCREAMING_SNAKE_CASE ) writer.writerow(['Movie title', 'IMDb rating'] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
254
1
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __lowerCAmelCase = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece class __a ( __UpperCamelCase , unittest.TestCase ): __lowercase : Union[str, Any] = XLMProphetNetTokenizer __lowercase : int = False __lowercase : Any = True def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing lowercase__: List[str] = XLMProphetNetTokenizer(lowerCAmelCase__ , keep_accents=lowerCAmelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self ) -> int: '''simple docstring''' lowercase__: Any = '[PAD]' lowercase__: Dict = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCAmelCase__ ) , lowerCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCAmelCase__ ) , lowerCAmelCase__ ) def SCREAMING_SNAKE_CASE__ ( self ) -> Union[str, Any]: '''simple docstring''' lowercase__: Optional[int] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '[PAD]' ) self.assertEqual(vocab_keys[1] , '[CLS]' ) self.assertEqual(vocab_keys[-1] , 'j' ) self.assertEqual(len(lowerCAmelCase__ ) , 1_012 ) def SCREAMING_SNAKE_CASE__ ( self ) -> int: '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1_012 ) def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' lowercase__: int = XLMProphetNetTokenizer(lowerCAmelCase__ , keep_accents=lowerCAmelCase__ ) lowercase__: Tuple = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCAmelCase__ , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) lowercase__: Tuple = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) lowercase__: Any = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, -9, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, -9, 4] ] , ) lowercase__: str = tokenizer.convert_ids_to_tokens(lowerCAmelCase__ ) self.assertListEqual( lowerCAmelCase__ , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '[UNK]', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '[UNK]', '.', ] , ) @cached_property def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' return XLMProphetNetTokenizer.from_pretrained('microsoft/xprophetnet-large-wiki100-cased' ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> List[Any]: '''simple docstring''' lowercase__: Union[str, Any] = 'Hello World!' lowercase__: Union[str, Any] = [35_389, 6_672, 49, 2] self.assertListEqual(lowerCAmelCase__ , self.big_tokenizer.encode(lowerCAmelCase__ ) ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' lowercase__: str = {'input_ids': [[11_073, 82_783, 18, 26, 82_783, 549, 51_540, 248, 17_209, 1_301, 217, 20, 215_186, 1_325, 147, 17_209, 1_301, 217, 20, 56_370, 53, 122_020, 20, 16_477, 27, 87_355, 4_548, 20, 4_728, 78_392, 17, 159_969, 18, 26, 24_491, 629, 15, 538, 22_704, 5_439, 15, 2_788, 24_491, 9_885, 15, 43_534, 605, 15, 814, 18_403, 33_200, 29, 15, 43_534, 24_458, 12_410, 111, 24_966, 83_669, 9_637, 144_068, 26, 850, 22_346, 27, 147, 24_966, 83_669, 83_490, 26, 39_113, 735, 27, 689, 656, 2_800, 1_339, 4_600, 53, 122_020, 115_785, 34, 816, 1_339, 46_887, 18, 147, 53_905, 1_951, 42_238, 41_170, 17_732, 834, 436, 15, 27_523, 98_733, 217, 147, 5_542, 4_981, 930, 17_347, 16, 2], [20_091, 629, 94, 82_786, 58, 490, 20, 1_528, 84, 53_905, 344, 80_592, 110_128, 18_822, 5_267, 1_306, 62, 152_537, 308, 7_997, 401, 124_427, 549, 35_442, 225, 109, 15_055, 25_748, 147, 7_119, 43_712, 34, 767, 135_366, 18, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [592, 63_784, 119_466, 17, 147_808, 88_214, 18, 656, 81, 32, 3_296, 10_280, 16, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCAmelCase__ , model_name='microsoft/xprophetnet-large-wiki100-cased' , revision='1acad1643ddd54a44df6a1b797ada8373685d90e' , )
196
_lowerCamelCase : int = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _lowerCamelCase : str = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] _lowerCamelCase : List[str] = { 0: '''Sunday''', 1: '''Monday''', 2: '''Tuesday''', 3: '''Wednesday''', 4: '''Thursday''', 5: '''Friday''', 6: '''Saturday''', } def a_ ( __lowercase : int , __lowercase : int , __lowercase : int ) -> str: assert len(str(__lowercase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: _snake_case = year // 100 _snake_case = (5 * (century % 4) + 2) % 7 _snake_case = year % 100 _snake_case = centurian % 12 _snake_case = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 _snake_case = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) _snake_case = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
282
0
"""simple docstring""" import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' @property def _UpperCAmelCase ( self ) -> Optional[Any]: torch.manual_seed(0 ) _a = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('''DownBlock2D''', '''AttnDownBlock2D''') , up_block_types=('''AttnUpBlock2D''', '''UpBlock2D''') , ) return model @property def _UpperCAmelCase ( self ) -> str: torch.manual_seed(0 ) _a = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=3 , ) return model @property def _UpperCAmelCase ( self ) -> Dict: torch.manual_seed(0 ) _a = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModel(_SCREAMING_SNAKE_CASE ) def _UpperCAmelCase ( self ) -> List[Any]: _a = self.dummy_uncond_unet _a = DDIMScheduler() _a = self.dummy_vq_model _a = LDMPipeline(unet=_SCREAMING_SNAKE_CASE , vqvae=_SCREAMING_SNAKE_CASE , scheduler=_SCREAMING_SNAKE_CASE ) ldm.to(_SCREAMING_SNAKE_CASE ) ldm.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) _a = torch.manual_seed(0 ) _a = ldm(generator=_SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type='''numpy''' ).images _a = torch.manual_seed(0 ) _a = ldm(generator=_SCREAMING_SNAKE_CASE , num_inference_steps=2 , output_type='''numpy''' , return_dict=_SCREAMING_SNAKE_CASE )[0] _a = image[0, -3:, -3:, -1] _a = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) _a = 1e-2 if torch_device != '''mps''' else 3e-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def _UpperCAmelCase ( self ) -> Optional[Any]: _a = LDMPipeline.from_pretrained('''CompVis/ldm-celebahq-256''' ) ldm.to(_SCREAMING_SNAKE_CASE ) ldm.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) _a = torch.manual_seed(0 ) _a = ldm(generator=_SCREAMING_SNAKE_CASE , num_inference_steps=5 , output_type='''numpy''' ).images _a = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a = np.array([0.4399, 0.44975, 0.46825, 0.474, 0.4359, 0.4581, 0.45095, 0.4341, 0.4447] ) _a = 1e-2 if torch_device != '''mps''' else 3e-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
368
"""simple docstring""" from collections.abc import Callable import numpy as np def A_ ( _lowerCAmelCase : Callable, _lowerCAmelCase : float, _lowerCAmelCase : float, _lowerCAmelCase : float, _lowerCAmelCase : float ): """simple docstring""" _a = int(np.ceil((x_end - xa) / step_size ) ) _a = np.zeros((n + 1,) ) _a = ya _a = xa for k in range(_lowerCAmelCase ): _a = y[k] + step_size * ode_func(_lowerCAmelCase, y[k] ) _a = y[k] + ( (step_size / 2) * (ode_func(_lowerCAmelCase, y[k] ) + ode_func(x + step_size, _lowerCAmelCase )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
153
0
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class A : __magic_name__ = 42 # setable values __magic_name__ = 42 __magic_name__ = 42 __magic_name__ = None @classmethod def __lowerCAmelCase ( cls , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" return cls(common=_UpperCamelCase , init_noise_sigma=_UpperCamelCase , timesteps=_UpperCamelCase ) @dataclass class A ( _lowerCamelCase ): __magic_name__ = 42 class A ( _lowerCamelCase , _lowerCamelCase ): __magic_name__ = [e.name for e in FlaxKarrasDiffusionSchedulers] __magic_name__ = 42 @property def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" return True @register_to_config def __init__( self , SCREAMING_SNAKE_CASE = 1000 , SCREAMING_SNAKE_CASE = 0.0_001 , SCREAMING_SNAKE_CASE = 0.02 , SCREAMING_SNAKE_CASE = "linear" , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = "fixed_small" , SCREAMING_SNAKE_CASE = True , SCREAMING_SNAKE_CASE = "epsilon" , SCREAMING_SNAKE_CASE = jnp.floataa , ) -> List[Any]: """simple docstring""" A : Tuple = dtype def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE = None ) -> DDPMSchedulerState: """simple docstring""" if common is None: A : int = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution A : Any = jnp.array(1.0 , dtype=self.dtype ) A : List[Any] = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_UpperCamelCase , init_noise_sigma=_UpperCamelCase , timesteps=_UpperCamelCase , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> jnp.ndarray: """simple docstring""" return sample def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = () ) -> DDPMSchedulerState: """simple docstring""" A : int = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 A : int = (jnp.arange(0 , _UpperCamelCase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_UpperCamelCase , timesteps=_UpperCamelCase , ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None ) -> Optional[Any]: """simple docstring""" A : List[str] = state.common.alphas_cumprod[t] A : Optional[Any] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample A : Any = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: A : Any = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": A : List[str] = jnp.clip(_UpperCamelCase , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": A : Tuple = jnp.log(jnp.clip(_UpperCamelCase , a_min=1e-20 ) ) elif variance_type == "fixed_large": A : str = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log A : Dict = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": A : Any = variance A : Optional[Any] = state.common.betas[t] A : Dict = (predicted_variance + 1) / 2 A : Any = frac * max_log + (1 - frac) * min_log return variance def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: """simple docstring""" A : str = timestep if key is None: A : Optional[Any] = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: A, A : Tuple = jnp.split(_UpperCamelCase , sample.shape[1] , axis=1 ) else: A : List[str] = None # 1. compute alphas, betas A : Dict = state.common.alphas_cumprod[t] A : Optional[Any] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) A : List[str] = 1 - alpha_prod_t A : Tuple = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": A : Union[str, Any] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": A : List[str] = model_output elif self.config.prediction_type == "v_prediction": A : Any = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F'prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ' ''' for the FlaxDDPMScheduler.''' ) # 3. Clip "predicted x_0" if self.config.clip_sample: A : Dict = jnp.clip(_UpperCamelCase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf A : Tuple = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t A : Optional[Any] = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf A : str = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): A : str = jax.random.split(_UpperCamelCase , num=1 ) A : Any = jax.random.normal(_UpperCamelCase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_UpperCamelCase , _UpperCamelCase , predicted_variance=_UpperCamelCase ) ** 0.5) * noise A : Any = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) A : Dict = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_UpperCamelCase , state=_UpperCamelCase ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> jnp.ndarray: """simple docstring""" return add_noise_common(state.common , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ) -> jnp.ndarray: """simple docstring""" return get_velocity_common(state.common , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) def __len__( self ) -> Any: """simple docstring""" return self.config.num_train_timesteps
3
import heapq import sys import numpy as np snake_case__ : Tuple = tuple[int, int] class A_ : def __init__(self :Union[str, Any] )-> Union[str, Any]: __A = [] __A = set() def _lowerCAmelCase (self :List[str] )-> str: if not self.empty(): return self.elements[0][0] else: return float('''inf''' ) def _lowerCAmelCase (self :Any )-> Any: return len(self.elements ) == 0 def _lowerCAmelCase (self :List[str] , _UpperCamelCase :Optional[Any] , _UpperCamelCase :Optional[int] )-> Optional[int]: if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(_UpperCamelCase ) else: # update # print("update", item) __A = [] ((__A) , (__A)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((__A) , (__A)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def _lowerCAmelCase (self :int , _UpperCamelCase :int )-> int: if item in self.set: self.set.remove(_UpperCamelCase ) __A = [] ((__A) , (__A)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((__A) , (__A)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def _lowerCAmelCase (self :Optional[Any] )-> int: return self.elements[0][1] def _lowerCAmelCase (self :List[str] )-> List[Any]: ((__A) , (__A)) = heapq.heappop(self.elements ) self.set.remove(_UpperCamelCase ) return (priority, item) def _a ( lowerCamelCase: TPos , lowerCamelCase: TPos ) -> Any: '''simple docstring''' __A = np.array(lowerCamelCase ) __A = np.array(lowerCamelCase ) return np.linalg.norm(a - b ) def _a ( lowerCamelCase: TPos , lowerCamelCase: TPos ) -> List[str]: '''simple docstring''' return consistent_heuristic(lowerCamelCase , lowerCamelCase ) // t def _a ( lowerCamelCase: TPos , lowerCamelCase: TPos ) -> Any: '''simple docstring''' return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def _a ( lowerCamelCase: TPos , lowerCamelCase: int , lowerCamelCase: TPos , lowerCamelCase: dict[TPos, float] ) -> Tuple: '''simple docstring''' __A = g_function[start] + Wa * heuristics[i](lowerCamelCase , lowerCamelCase ) return ans def _a ( lowerCamelCase: Tuple , lowerCamelCase: Optional[int] , lowerCamelCase: Optional[int] ) -> Tuple: '''simple docstring''' __A = np.chararray((n, n) ) for i in range(lowerCamelCase ): for j in range(lowerCamelCase ): __A = '''*''' for i in range(lowerCamelCase ): for j in range(lowerCamelCase ): if (j, (n - 1) - i) in blocks: __A = '''#''' __A = '''-''' __A = back_pointer[goal] while x != start: ((__A) , (__A)) = x # print(x) __A = '''-''' __A = back_pointer[x] __A = '''-''' for i in range(lowerCamelCase ): for j in range(lowerCamelCase ): if (i, j) == (0, n - 1): print(grid[i][j] , end=''' ''' ) print('''<-- End position''' , end=''' ''' ) else: print(grid[i][j] , end=''' ''' ) print() print('''^''' ) print('''Start position''' ) print() print('''# is an obstacle''' ) print('''- is the path taken by algorithm''' ) print('''PATH TAKEN BY THE ALGORITHM IS:-''' ) __A = back_pointer[goal] while x != start: print(lowerCamelCase , end=''' ''' ) __A = back_pointer[x] print(lowerCamelCase ) sys.exit() def _a ( lowerCamelCase: TPos ) -> Optional[Any]: '''simple docstring''' if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def _a ( lowerCamelCase: int , lowerCamelCase: Optional[int] , lowerCamelCase: Tuple , lowerCamelCase: Tuple , lowerCamelCase: Optional[int] , lowerCamelCase: str , lowerCamelCase: List[Any] , lowerCamelCase: Dict , ) -> Dict: '''simple docstring''' for itera in range(lowerCamelCase ): open_list[itera].remove_element(lowerCamelCase ) # print("s", s) # print("j", j) ((__A) , (__A)) = s __A = (x - 1, y) __A = (x + 1, y) __A = (x, y + 1) __A = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(lowerCamelCase ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(lowerCamelCase ) __A = -1 __A = float('''inf''' ) if valid(lowerCamelCase ) and g_function[neighbours] > g_function[s] + 1: __A = g_function[s] + 1 __A = s if neighbours not in close_list_anchor: open_list[0].put(lowerCamelCase , key(lowerCamelCase , 0 , lowerCamelCase , lowerCamelCase ) ) if neighbours not in close_list_inad: for var in range(1 , lowerCamelCase ): if key(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) <= Wa * key( lowerCamelCase , 0 , lowerCamelCase , lowerCamelCase ): open_list[j].put( lowerCamelCase , key(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) ) def _a ( ) -> str: '''simple docstring''' __A = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(15 , 20 ): some_list.append((x, 17) ) for x in range(10 , 19 ): for y in range(1 , 15 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(12 , 19 ): some_list.append((x, y) ) for x in range(3 , 13 ): for y in range(16 , 19 ): some_list.append((x, y) ) return some_list snake_case__ : Union[str, Any] = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} snake_case__ : List[str] = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] snake_case__ : List[str] = make_common_ground() snake_case__ : int = blocks_blk # hyper parameters snake_case__ : Optional[int] = 1 snake_case__ : List[Any] = 1 snake_case__ : int = 20 snake_case__ : Optional[int] = 3 # one consistent and two other inconsistent # start and end destination snake_case__ : Tuple = (0, 0) snake_case__ : Any = (n - 1, n - 1) snake_case__ : List[Any] = 1 def _a ( lowerCamelCase: TPos , lowerCamelCase: TPos , lowerCamelCase: int ) -> Union[str, Any]: '''simple docstring''' __A = {start: 0, goal: float('''inf''' )} __A = {start: -1, goal: -1} __A = [] __A = set() for i in range(lowerCamelCase ): open_list.append(PriorityQueue() ) open_list[i].put(lowerCamelCase , key(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) ) __A = [] __A = [] while open_list[0].minkey() < float('''inf''' ): for i in range(1 , lowerCamelCase ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float('''inf''' ): do_something(lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: __A , __A = open_list[i].top_show() visited.add(lowerCamelCase ) expand_state( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) close_list_inad.append(lowerCamelCase ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float('''inf''' ): do_something(lowerCamelCase , lowerCamelCase , lowerCamelCase ) else: __A = open_list[0].top_show() visited.add(lowerCamelCase ) expand_state( lowerCamelCase , 0 , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) close_list_anchor.append(lowerCamelCase ) print('''No path found to goal''' ) print() for i in range(n - 1 , -1 , -1 ): for j in range(lowerCamelCase ): if (j, i) in blocks: print('''#''' , end=''' ''' ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print('''*''' , end=''' ''' ) else: print('''-''' , end=''' ''' ) else: print('''*''' , end=''' ''' ) if (j, i) == (n - 1, n - 1): print('''<-- End position''' , end=''' ''' ) print() print('''^''' ) print('''Start position''' ) print() print('''# is an obstacle''' ) print('''- is the path taken by algorithm''' ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
117
0
import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __lowerCAmelCase ( lowerCAmelCase , unittest.TestCase): _a = FunnelTokenizer _a = FunnelTokenizerFast _a = True _a = True def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): super().setUp() lowercase :int = [ "<unk>", "<cls>", "<sep>", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] lowercase :Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def SCREAMING_SNAKE_CASE ( self: Any , **_lowerCAmelCase: List[str] ): return FunnelTokenizer.from_pretrained(self.tmpdirname , **__A ) def SCREAMING_SNAKE_CASE ( self: Any , **_lowerCAmelCase: List[Any] ): return FunnelTokenizerFast.from_pretrained(self.tmpdirname , **__A ) def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: Tuple ): lowercase :Optional[Any] = "UNwant\u00E9d,running" lowercase :Tuple = "unwanted, running" return input_text, output_text def SCREAMING_SNAKE_CASE ( self: Any ): lowercase :List[Any] = self.tokenizer_class(self.vocab_file ) lowercase :int = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(__A , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__A ) , [7, 4, 5, 10, 8, 9] ) def SCREAMING_SNAKE_CASE ( self: Optional[Any] ): lowercase :Union[str, Any] = self.get_tokenizers(do_lower_case=__A ) for tokenizer in tokenizers: lowercase :Optional[int] = tokenizer("UNwant\u00E9d,running" ) lowercase :Optional[int] = len(inputs["input_ids"] ) - 1 self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len ) lowercase :str = tokenizer("UNwant\u00E9d,running" , "UNwant\u00E9d,running" ) self.assertListEqual(inputs["token_type_ids"] , [2] + [0] * sentence_len + [1] * sentence_len )
359
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, **lowerCamelCase ): lowercase :List[Any] = AutoConfig.from_pretrained(lowerCamelCase, **lowerCamelCase ) lowercase :Union[str, Any] = AutoModelForSeqaSeqLM.from_config(lowerCamelCase ) model.save_pretrained(lowerCamelCase ) AutoTokenizer.from_pretrained(lowerCamelCase ).save_pretrained(lowerCamelCase ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
158
0
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class A__ ( __snake_case ): def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(A_ , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(A_ , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(A_ , "num_attention_heads" ) ) class A__ : def __init__( self , A_ , A_=13 , A_=32 , A_=2 , A_=3 , A_=640 , A_=4 , A_="silu" , A_=3 , A_=32 , A_=0.1 , A_=0.1 , A_=0.1 , A_=0.02 , A_=True , A_=True , A_=10 , A_=None , ): '''simple docstring''' UpperCamelCase : Optional[Any] = parent UpperCamelCase : List[Any] = batch_size UpperCamelCase : Optional[int] = image_size UpperCamelCase : List[Any] = patch_size UpperCamelCase : Any = num_channels UpperCamelCase : Optional[int] = last_hidden_size UpperCamelCase : str = num_attention_heads UpperCamelCase : Optional[int] = hidden_act UpperCamelCase : Dict = conv_kernel_size UpperCamelCase : Any = output_stride UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : Optional[Any] = attention_probs_dropout_prob UpperCamelCase : Dict = classifier_dropout_prob UpperCamelCase : Optional[int] = use_labels UpperCamelCase : str = is_training UpperCamelCase : Tuple = num_labels UpperCamelCase : Optional[Any] = initializer_range UpperCamelCase : int = scope def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase : List[Any] = None UpperCamelCase : Optional[int] = None if self.use_labels: UpperCamelCase : List[Any] = ids_tensor([self.batch_size] , self.num_labels ) UpperCamelCase : List[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) UpperCamelCase : List[str] = self.get_config() return config, pixel_values, labels, pixel_labels def __UpperCamelCase( self ): '''simple docstring''' return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Optional[int] = MobileViTModel(config=A_ ) model.to(A_ ) model.eval() UpperCamelCase : Tuple = model(A_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : Tuple = self.num_labels UpperCamelCase : int = MobileViTForImageClassification(A_ ) model.to(A_ ) model.eval() UpperCamelCase : Union[str, Any] = model(A_ , labels=A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __UpperCamelCase( self , A_ , A_ , A_ , A_ ): '''simple docstring''' UpperCamelCase : List[Any] = self.num_labels UpperCamelCase : Any = MobileViTForSemanticSegmentation(A_ ) model.to(A_ ) model.eval() UpperCamelCase : Optional[int] = model(A_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) UpperCamelCase : str = model(A_ , labels=A_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase : List[str] = config_and_inputs UpperCamelCase : Union[str, Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A__ ( __snake_case , __snake_case , unittest.TestCase ): _UpperCAmelCase :Union[str, Any] = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) _UpperCAmelCase :Dict = ( { 'feature-extraction': MobileViTModel, 'image-classification': MobileViTForImageClassification, 'image-segmentation': MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) _UpperCAmelCase :Tuple = False _UpperCAmelCase :int = False _UpperCAmelCase :Tuple = False _UpperCAmelCase :Union[str, Any] = False def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Any = MobileViTModelTester(self ) UpperCamelCase : int = MobileViTConfigTester(self , config_class=A_ , has_text_modality=A_ ) def __UpperCamelCase( self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __UpperCamelCase( self ): '''simple docstring''' pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __UpperCamelCase( self ): '''simple docstring''' pass @unittest.skip(reason="MobileViT does not output attentions" ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase , UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Optional[Any] = model_class(A_ ) UpperCamelCase : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase : Optional[int] = [*signature.parameters.keys()] UpperCamelCase : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , A_ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __UpperCamelCase( self ): '''simple docstring''' pass def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' def check_hidden_states_output(A_ , A_ , A_ ): UpperCamelCase : Tuple = model_class(A_ ) model.to(A_ ) model.eval() with torch.no_grad(): UpperCamelCase : Dict = model(**self._prepare_for_class(A_ , A_ ) ) UpperCamelCase : str = outputs.hidden_states UpperCamelCase : Optional[int] = 5 self.assertEqual(len(A_ ) , A_ ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. UpperCamelCase : Tuple = 2 for i in range(len(A_ ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) UpperCamelCase , UpperCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase : Union[str, Any] = True check_hidden_states_output(A_ , A_ , A_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCamelCase : int = True check_hidden_states_output(A_ , A_ , A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A_ ) def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A_ ) @slow def __UpperCamelCase( self ): '''simple docstring''' for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : Tuple = MobileViTModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) def A_ ( ) -> Union[str, Any]: UpperCamelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A__ ( unittest.TestCase ): @cached_property def __UpperCamelCase( self ): '''simple docstring''' return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : List[str] = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(A_ ) UpperCamelCase : List[Any] = self.default_image_processor UpperCamelCase : Any = prepare_img() UpperCamelCase : Union[str, Any] = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : List[str] = model(**A_ ) # verify the logits UpperCamelCase : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , A_ ) UpperCamelCase : Any = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(A_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , A_ , atol=1e-4 ) ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : Tuple = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) UpperCamelCase : str = model.to(A_ ) UpperCamelCase : Optional[int] = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) UpperCamelCase : Tuple = prepare_img() UpperCamelCase : Dict = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : Dict = model(**A_ ) UpperCamelCase : str = outputs.logits # verify the logits UpperCamelCase : int = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , A_ ) UpperCamelCase : Any = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=A_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , A_ , atol=1e-4 ) ) @slow def __UpperCamelCase( self ): '''simple docstring''' UpperCamelCase : int = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) UpperCamelCase : Any = model.to(A_ ) UpperCamelCase : List[Any] = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) UpperCamelCase : Union[str, Any] = prepare_img() UpperCamelCase : Dict = image_processor(images=A_ , return_tensors="pt" ).to(A_ ) # forward pass with torch.no_grad(): UpperCamelCase : Tuple = model(**A_ ) UpperCamelCase : Tuple = outputs.logits.detach().cpu() UpperCamelCase : int = image_processor.post_process_semantic_segmentation(outputs=A_ , target_sizes=[(50, 60)] ) UpperCamelCase : Optional[int] = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , A_ ) UpperCamelCase : List[str] = image_processor.post_process_semantic_segmentation(outputs=A_ ) UpperCamelCase : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , A_ )
52
def A_ ( _lowerCAmelCase ) -> str: UpperCamelCase : List[Any] = "" for ch in key: if ch == " " or ch not in key_no_dups and ch.isalpha(): key_no_dups += ch return key_no_dups def A_ ( _lowerCAmelCase ) -> dict[str, str]: UpperCamelCase : Optional[Any] = [chr(i + 65 ) for i in range(26 )] # Remove duplicate characters from key UpperCamelCase : Tuple = remove_duplicates(key.upper() ) UpperCamelCase : int = len(_lowerCAmelCase ) # First fill cipher with key characters UpperCamelCase : int = {alphabet[i]: char for i, char in enumerate(_lowerCAmelCase )} # Then map remaining characters in alphabet to # the alphabet from the beginning for i in range(len(_lowerCAmelCase ) , 26 ): UpperCamelCase : Optional[Any] = alphabet[i - offset] # Ensure we are not mapping letters to letters previously mapped while char in key: offset -= 1 UpperCamelCase : List[str] = alphabet[i - offset] UpperCamelCase : List[Any] = char return cipher_alphabet def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> str: return "".join(cipher_map.get(_lowerCAmelCase , _lowerCAmelCase ) for ch in message.upper() ) def A_ ( _lowerCAmelCase , _lowerCAmelCase ) -> str: UpperCamelCase : Union[str, Any] = {v: k for k, v in cipher_map.items()} return "".join(rev_cipher_map.get(_lowerCAmelCase , _lowerCAmelCase ) for ch in message.upper() ) def A_ ( ) -> None: UpperCamelCase : int = input("Enter message to encode or decode: " ).strip() UpperCamelCase : str = input("Enter keyword: " ).strip() UpperCamelCase : Union[str, Any] = input("Encipher or decipher? E/D:" ).strip()[0].lower() try: UpperCamelCase : List[str] = {"e": encipher, "d": decipher}[option] except KeyError: raise KeyError("invalid input option" ) UpperCamelCase : str = create_cipher_map(_lowerCAmelCase ) print(func(_lowerCAmelCase , _lowerCAmelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
52
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase : int = { "configuration_bert": ["BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BertConfig", "BertOnnxConfig"], "tokenization_bert": ["BasicTokenizer", "BertTokenizer", "WordpieceTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[Any] = ["BertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : str = [ "BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "BertForMaskedLM", "BertForMultipleChoice", "BertForNextSentencePrediction", "BertForPreTraining", "BertForQuestionAnswering", "BertForSequenceClassification", "BertForTokenClassification", "BertLayer", "BertLMHeadModel", "BertModel", "BertPreTrainedModel", "load_tf_weights_in_bert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[Any] = [ "TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFBertEmbeddings", "TFBertForMaskedLM", "TFBertForMultipleChoice", "TFBertForNextSentencePrediction", "TFBertForPreTraining", "TFBertForQuestionAnswering", "TFBertForSequenceClassification", "TFBertForTokenClassification", "TFBertLMHeadModel", "TFBertMainLayer", "TFBertModel", "TFBertPreTrainedModel", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : List[str] = ["TFBertTokenizer"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase : Any = [ "FlaxBertForCausalLM", "FlaxBertForMaskedLM", "FlaxBertForMultipleChoice", "FlaxBertForNextSentencePrediction", "FlaxBertForPreTraining", "FlaxBertForQuestionAnswering", "FlaxBertForSequenceClassification", "FlaxBertForTokenClassification", "FlaxBertModel", "FlaxBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys UpperCAmelCase : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
313
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase=False ) -> Union[str, Any]: '''simple docstring''' lowercase_ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F'''module.blocks.{i}.norm1.weight''', F'''vit.encoder.layer.{i}.layernorm_before.weight''') ) rename_keys.append((F'''module.blocks.{i}.norm1.bias''', F'''vit.encoder.layer.{i}.layernorm_before.bias''') ) rename_keys.append( (F'''module.blocks.{i}.attn.proj.weight''', F'''vit.encoder.layer.{i}.attention.output.dense.weight''') ) rename_keys.append((F'''module.blocks.{i}.attn.proj.bias''', F'''vit.encoder.layer.{i}.attention.output.dense.bias''') ) rename_keys.append((F'''module.blocks.{i}.norm2.weight''', F'''vit.encoder.layer.{i}.layernorm_after.weight''') ) rename_keys.append((F'''module.blocks.{i}.norm2.bias''', F'''vit.encoder.layer.{i}.layernorm_after.bias''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc1.weight''', F'''vit.encoder.layer.{i}.intermediate.dense.weight''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc1.bias''', F'''vit.encoder.layer.{i}.intermediate.dense.bias''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc2.weight''', F'''vit.encoder.layer.{i}.output.dense.weight''') ) rename_keys.append((F'''module.blocks.{i}.mlp.fc2.bias''', F'''vit.encoder.layer.{i}.output.dense.bias''') ) # projection layer + position embeddings rename_keys.extend( [ ("""module.cls_token""", """vit.embeddings.cls_token"""), ("""module.patch_embed.proj.weight""", """vit.embeddings.patch_embeddings.projection.weight"""), ("""module.patch_embed.proj.bias""", """vit.embeddings.patch_embeddings.projection.bias"""), ("""module.pos_embed""", """vit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""module.norm.weight""", """layernorm.weight"""), ("""module.norm.bias""", """layernorm.bias"""), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" lowercase_ = [(pair[0], pair[1][4:]) if pair[1].startswith("""vit""" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("""norm.weight""", """vit.layernorm.weight"""), ("""norm.bias""", """vit.layernorm.bias"""), ("""head.weight""", """classifier.weight"""), ("""head.bias""", """classifier.bias"""), ] ) return rename_keys def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> Any: '''simple docstring''' for i in range(config.num_hidden_layers ): if base_model: lowercase_ = """""" else: lowercase_ = """vit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowercase_ = state_dict.pop(F'''module.blocks.{i}.attn.qkv.weight''' ) lowercase_ = state_dict.pop(F'''module.blocks.{i}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict lowercase_ = in_proj_weight[ : config.hidden_size, : ] lowercase_ = in_proj_bias[: config.hidden_size] lowercase_ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase_ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowercase_ = in_proj_weight[ -config.hidden_size :, : ] lowercase_ = in_proj_bias[-config.hidden_size :] def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> List[str]: '''simple docstring''' lowercase_ = ["""head.weight""", """head.bias"""] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> str: '''simple docstring''' lowercase_ = [ """module.fc.fc1.weight""", """module.fc.fc1.bias""", """module.fc.bn1.weight""", """module.fc.bn1.bias""", """module.fc.bn1.running_mean""", """module.fc.bn1.running_var""", """module.fc.bn1.num_batches_tracked""", """module.fc.fc2.weight""", """module.fc.fc2.bias""", """module.fc.bn2.weight""", """module.fc.bn2.bias""", """module.fc.bn2.running_mean""", """module.fc.bn2.running_var""", """module.fc.bn2.num_batches_tracked""", """module.fc.fc3.weight""", """module.fc.fc3.bias""", ] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Tuple: '''simple docstring''' lowercase_ = dct.pop(__lowerCAmelCase ) lowercase_ = val def _SCREAMING_SNAKE_CASE (__lowerCAmelCase , __lowerCAmelCase ) -> Tuple: '''simple docstring''' lowercase_ = ViTMSNConfig() lowercase_ = 10_00 lowercase_ = """datasets/huggingface/label-files""" lowercase_ = """imagenet-1k-id2label.json""" lowercase_ = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase ) , """r""" ) ) lowercase_ = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} lowercase_ = idalabel lowercase_ = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: lowercase_ = 3_84 lowercase_ = 15_36 lowercase_ = 6 elif "l16" in checkpoint_url: lowercase_ = 10_24 lowercase_ = 40_96 lowercase_ = 24 lowercase_ = 16 lowercase_ = 0.1 elif "b4" in checkpoint_url: lowercase_ = 4 elif "l7" in checkpoint_url: lowercase_ = 7 lowercase_ = 10_24 lowercase_ = 40_96 lowercase_ = 24 lowercase_ = 16 lowercase_ = 0.1 lowercase_ = ViTMSNModel(__lowerCAmelCase ) lowercase_ = torch.hub.load_state_dict_from_url(__lowerCAmelCase , map_location="""cpu""" )["""target_encoder"""] lowercase_ = ViTImageProcessor(size=config.image_size ) remove_projection_head(__lowerCAmelCase ) lowercase_ = create_rename_keys(__lowerCAmelCase , base_model=__lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , base_model=__lowerCAmelCase ) model.load_state_dict(__lowerCAmelCase ) model.eval() lowercase_ = """http://images.cocodataset.org/val2017/000000039769.jpg""" lowercase_ = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) lowercase_ = ViTImageProcessor( size=config.image_size , image_mean=__lowerCAmelCase , image_std=__lowerCAmelCase ) lowercase_ = image_processor(images=__lowerCAmelCase , return_tensors="""pt""" ) # forward pass torch.manual_seed(2 ) lowercase_ = model(**__lowerCAmelCase ) lowercase_ = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: lowercase_ = torch.tensor([[-1.0_915, -1.4_876, -1.1_809]] ) elif "b16" in checkpoint_url: lowercase_ = torch.tensor([[14.2_889, -18.9_045, 11.7_281]] ) elif "l16" in checkpoint_url: lowercase_ = torch.tensor([[41.5_028, -22.8_681, 45.6_475]] ) elif "b4" in checkpoint_url: lowercase_ = torch.tensor([[-4.3_868, 5.2_932, -0.4_137]] ) else: lowercase_ = torch.tensor([[-0.1_792, -0.6_465, 2.4_263]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , __lowerCAmelCase , atol=1E-4 ) print(F'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(__lowerCAmelCase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": UpperCAmelCase : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar", type=str, help="URL of the checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) UpperCAmelCase : Tuple = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
313
1
"""simple docstring""" from __future__ import annotations def _A ( UpperCamelCase_ : list[int | str]) -> None: '''simple docstring''' create_state_space_tree(UpperCamelCase_, [], 0, [0 for i in range(len(UpperCamelCase_))]) def _A ( UpperCamelCase_ : list[int | str], UpperCamelCase_ : list[int | str], UpperCamelCase_ : int, UpperCamelCase_ : list[int], ) -> None: '''simple docstring''' if index == len(UpperCamelCase_): print(UpperCamelCase_) return for i in range(len(UpperCamelCase_)): if not index_used[i]: current_sequence.append(sequence[i]) __lowercase = True create_state_space_tree(UpperCamelCase_, UpperCamelCase_, index + 1, UpperCamelCase_) current_sequence.pop() __lowercase = False _a = [3, 1, 2, 4] generate_all_permutations(sequence) _a = ["A", "B", "C"] generate_all_permutations(sequence_a)
17
"""simple docstring""" import torch from transformers import AutoModel class SCREAMING_SNAKE_CASE ( torch.nn.Module ): """simple docstring""" def __init__( self : Tuple ,lowercase_ : Dict="sayef/fsner-bert-base-uncased" ): super(lowercase_ ,self ).__init__() lowerCAmelCase__ : int = AutoModel.from_pretrained(lowercase_ ,return_dict=lowercase_ ) lowerCAmelCase__ : Optional[int] = torch.nn.CosineSimilarity(3 ,1E-08 ) lowerCAmelCase__ : List[str] = torch.nn.Softmax(dim=1 ) def __lowerCAmelCase ( self : str ,**lowercase_ : int ): return self.bert(**lowercase_ ).last_hidden_state def __lowerCAmelCase ( self : List[Any] ,lowercase_ : Optional[int] ): return token_embeddings.sum(2 ,keepdim=lowercase_ ) def __lowerCAmelCase ( self : Dict ,lowercase_ : int ,lowercase_ : str ,lowercase_ : Tuple=1 ): return self.softmax(T * self.cos(lowercase_ ,lowercase_ ) ) def __lowerCAmelCase ( self : Optional[Any] ,lowercase_ : str ,lowercase_ : Union[str, Any] ): lowerCAmelCase__ : List[Any] = W_supports['''sizes'''].tolist() lowerCAmelCase__ : Dict = W_supports['''start_token_id'''].item() lowerCAmelCase__ : Union[str, Any] = W_supports['''end_token_id'''].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] lowerCAmelCase__ : Optional[Any] = self.BERT(**lowercase_ ) lowerCAmelCase__ : int = self.BERT(**lowercase_ ) lowerCAmelCase__ : List[str] = None lowerCAmelCase__ : Union[str, Any] = None lowerCAmelCase__ : int = W_supports['''input_ids'''] == start_token_id lowerCAmelCase__ : Optional[Any] = W_supports['''input_ids'''] == end_token_id for i, size in enumerate(lowercase_ ): if i == 0: lowerCAmelCase__ : str = 0 else: lowerCAmelCase__ : List[Any] = support_sizes[i - 1] lowerCAmelCase__ : Optional[Any] = S[s : s + size][start_token_masks[s : s + size]] lowerCAmelCase__ : List[Any] = S[s : s + size][end_token_masks[s : s + size]] lowerCAmelCase__ : Union[str, Any] = torch.matmul(q[i] ,s_start.T ).sum(1 ).softmax(0 ) lowerCAmelCase__ : Any = torch.matmul(q[i] ,s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: lowerCAmelCase__ : List[Any] = torch.vstack((p_starts, p_start) ) lowerCAmelCase__ : List[Any] = torch.vstack((p_ends, p_end) ) else: lowerCAmelCase__ : Union[str, Any] = p_start lowerCAmelCase__ : str = p_end return p_starts, p_ends
106
0
"""simple docstring""" def __a ( ) ->Tuple: a__: str = [] a__: Any = 1 while len(_SCREAMING_SNAKE_CASE ) < 1e6: constant.append(str(_SCREAMING_SNAKE_CASE ) ) i += 1 a__: int = ''.join(_SCREAMING_SNAKE_CASE ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[99999] ) * int(constant[999999] ) ) if __name__ == "__main__": print(solution())
356
"""simple docstring""" import gc import inspect import unittest import torch from parameterized import parameterized from diffusers import PriorTransformer from diffusers.utils import floats_tensor, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin enable_full_determinism() class __snake_case ( __lowerCAmelCase , unittest.TestCase ): a__ = PriorTransformer a__ = """hidden_states""" @property def lowerCamelCase_ ( self) -> Tuple: '''simple docstring''' a__: Union[str, Any] = 4 a__: Any = 8 a__: Optional[Any] = 7 a__: Tuple = floats_tensor((batch_size, embedding_dim)).to(lowercase) a__: Optional[int] = floats_tensor((batch_size, embedding_dim)).to(lowercase) a__: List[str] = floats_tensor((batch_size, num_embeddings, embedding_dim)).to(lowercase) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def lowerCamelCase_ ( self , lowercase=0) -> str: '''simple docstring''' torch.manual_seed(lowercase) a__: Optional[Any] = 4 a__: Optional[Any] = 8 a__: Union[str, Any] = 7 a__: Optional[Any] = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: List[str] = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: Tuple = torch.randn((batch_size, num_embeddings, embedding_dim)).to(lowercase) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } @property def lowerCamelCase_ ( self) -> str: '''simple docstring''' return (4, 8) @property def lowerCamelCase_ ( self) -> Optional[int]: '''simple docstring''' return (4, 8) def lowerCamelCase_ ( self) -> str: '''simple docstring''' a__: int = { 'num_attention_heads': 2, 'attention_head_dim': 4, 'num_layers': 2, 'embedding_dim': 8, 'num_embeddings': 7, 'additional_embeddings': 4, } a__: Union[str, Any] = self.dummy_input return init_dict, inputs_dict def lowerCamelCase_ ( self) -> Union[str, Any]: '''simple docstring''' a__ , a__: Union[str, Any] = PriorTransformer.from_pretrained( 'hf-internal-testing/prior-dummy' , output_loading_info=lowercase) self.assertIsNotNone(lowercase) self.assertEqual(len(loading_info['missing_keys']) , 0) model.to(lowercase) a__: Any = model(**self.dummy_input)[0] assert hidden_states is not None, "Make sure output is not None" def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__ , a__: Tuple = self.prepare_init_args_and_inputs_for_common() a__: Any = self.model_class(**lowercase) a__: str = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__: Tuple = [*signature.parameters.keys()] a__: List[Any] = ['hidden_states', 'timestep'] self.assertListEqual(arg_names[:2] , lowercase) def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' a__: str = PriorTransformer.from_pretrained('hf-internal-testing/prior-dummy') a__: str = model.to(lowercase) if hasattr(lowercase , 'set_default_attn_processor'): model.set_default_attn_processor() a__: Dict = self.get_dummy_seed_input() with torch.no_grad(): a__: str = model(**lowercase)[0] a__: str = output[0, :5].flatten().cpu() print(lowercase) # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. a__: Any = torch.tensor([-1.3436, -0.2870, 0.7538, 0.4368, -0.0239]) self.assertTrue(torch_all_close(lowercase , lowercase , rtol=1e-2)) @slow class __snake_case ( unittest.TestCase ): def lowerCamelCase_ ( self , lowercase=1 , lowercase=7_68 , lowercase=77 , lowercase=0) -> int: '''simple docstring''' torch.manual_seed(lowercase) a__: Union[str, Any] = batch_size a__: List[str] = embedding_dim a__: str = num_embeddings a__: Tuple = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: List[str] = torch.randn((batch_size, embedding_dim)).to(lowercase) a__: str = torch.randn((batch_size, num_embeddings, embedding_dim)).to(lowercase) return { "hidden_states": hidden_states, "timestep": 2, "proj_embedding": proj_embedding, "encoder_hidden_states": encoder_hidden_states, } def lowerCamelCase_ ( self) -> List[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() @parameterized.expand( [ # fmt: off [13, [-0.5861, 0.1283, -0.0931, 0.0882, 0.4476, 0.1329, -0.0498, 0.0640]], [37, [-0.4913, 0.0110, -0.0483, 0.0541, 0.4954, -0.0170, 0.0354, 0.1651]], # fmt: on ]) def lowerCamelCase_ ( self , lowercase , lowercase) -> str: '''simple docstring''' a__: Tuple = PriorTransformer.from_pretrained('kandinsky-community/kandinsky-2-1-prior' , subfolder='prior') model.to(lowercase) a__: Optional[Any] = self.get_dummy_seed_input(seed=lowercase) with torch.no_grad(): a__: Optional[int] = model(**lowercase)[0] assert list(sample.shape) == [1, 7_68] a__: List[str] = sample[0, :8].flatten().cpu() print(lowercase) a__: Union[str, Any] = torch.tensor(lowercase) assert torch_all_close(lowercase , lowercase , atol=1e-3)
203
0
"""simple docstring""" from PIL import Image def a__ ( snake_case__ , snake_case__ ) -> Image: lowerCamelCase = (2_59 * (level + 2_55)) / (2_55 * (2_59 - level)) def contrast(snake_case__ ) -> int: return int(1_28 + factor * (c - 1_28) ) return img.point(snake_case__ ) if __name__ == "__main__": # Load image with Image.open("""image_data/lena.jpg""") as img: # Change contrast to 170 lowerCAmelCase : Optional[Any] = change_contrast(img, 170) cont_img.save("""image_data/lena_high_contrast.png""", format="""png""")
291
"""simple docstring""" from math import ceil def a__ ( snake_case__ , snake_case__ ) -> Optional[int]: lowerCamelCase = list(range(0 , snake_case__ ) ) lowerCamelCase = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check lowerCamelCase = [] for i in device_map_blocks: if device_map_blocks.count(snake_case__ ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(snake_case__ ) # Missing blocks lowerCamelCase = [i for i in blocks if i not in device_map_blocks] lowerCamelCase = [i for i in device_map_blocks if i not in blocks] if len(snake_case__ ) != 0: raise ValueError( """Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device.""" """ These attention blocks were specified more than once: """ + str(snake_case__ ) ) if len(snake_case__ ) != 0: raise ValueError( """There are attention blocks for this model that are not specified in the device_map. Add these attention """ """blocks to a device on the device_map: """ + str(snake_case__ ) ) if len(snake_case__ ) != 0: raise ValueError( """The device_map contains more attention blocks than this model has. Remove these from the device_map:""" + str(snake_case__ ) ) def a__ ( snake_case__ , snake_case__ ) -> List[Any]: lowerCamelCase = list(range(snake_case__ ) ) lowerCamelCase = int(ceil(n_layers / len(snake_case__ ) ) ) lowerCamelCase = [layers[i : i + n_blocks] for i in range(0 , snake_case__ , snake_case__ )] return dict(zip(snake_case__ , snake_case__ ) )
291
1
def _A ( lowerCAmelCase_ : int ): """simple docstring""" return str(lowerCAmelCase_ ) == str(lowerCAmelCase_ )[::-1] def _A ( lowerCAmelCase_ : int ): """simple docstring""" return int(lowerCAmelCase_ ) + int(str(lowerCAmelCase_ )[::-1] ) def _A ( lowerCAmelCase_ : int = 1_0000 ): """simple docstring""" lowerCAmelCase__ = [] for num in range(1 , lowerCAmelCase_ ): lowerCAmelCase__ = 0 lowerCAmelCase__ = num while iterations < 50: lowerCAmelCase__ = sum_reverse(lowerCAmelCase_ ) iterations += 1 if is_palindrome(lowerCAmelCase_ ): break else: lychrel_nums.append(lowerCAmelCase_ ) return len(lowerCAmelCase_ ) if __name__ == "__main__": print(F"""{solution() = }""")
221
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def _A ( lowerCAmelCase_ : str , lowerCAmelCase_ : str , **lowerCAmelCase_ : str ): """simple docstring""" lowerCAmelCase__ = AutoConfig.from_pretrained(lowerCAmelCase_ , **lowerCAmelCase_ ) lowerCAmelCase__ = AutoModelForSeqaSeqLM.from_config(lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) AutoTokenizer.from_pretrained(lowerCAmelCase_ ).save_pretrained(lowerCAmelCase_ ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
221
1
from ...configuration_utils import PretrainedConfig from ...utils import logging A__: Any = logging.get_logger(__name__) A__: List[str] = { '''google/pegasus-large''': '''https://huggingface.co/google/pegasus-large/resolve/main/config.json''', # See all PEGASUS models at https://huggingface.co/models?filter=pegasus } class _a ( UpperCamelCase__): """simple docstring""" UpperCamelCase__ = """pegasus""" UpperCamelCase__ = ["""past_key_values"""] UpperCamelCase__ = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self: List[str] , __lowerCamelCase: Dict=5_0265 , __lowerCamelCase: int=1024 , __lowerCamelCase: Dict=12 , __lowerCamelCase: Dict=4096 , __lowerCamelCase: str=16 , __lowerCamelCase: List[Any]=12 , __lowerCamelCase: int=4096 , __lowerCamelCase: Optional[Any]=16 , __lowerCamelCase: Union[str, Any]=0.0 , __lowerCamelCase: int=0.0 , __lowerCamelCase: List[str]=True , __lowerCamelCase: Union[str, Any]=True , __lowerCamelCase: List[Any]="gelu" , __lowerCamelCase: List[Any]=1024 , __lowerCamelCase: int=0.1 , __lowerCamelCase: Union[str, Any]=0.0 , __lowerCamelCase: Union[str, Any]=0.0 , __lowerCamelCase: Any=0.02 , __lowerCamelCase: Union[str, Any]=0 , __lowerCamelCase: str=False , __lowerCamelCase: Optional[int]=0 , __lowerCamelCase: Optional[Any]=1 , __lowerCamelCase: Optional[int]=1 , **__lowerCamelCase: Union[str, Any] , ): '''simple docstring''' UpperCamelCase__: Union[str, Any] = vocab_size UpperCamelCase__: List[Any] = max_position_embeddings UpperCamelCase__: Tuple = d_model UpperCamelCase__: str = encoder_ffn_dim UpperCamelCase__: Optional[int] = encoder_layers UpperCamelCase__: List[Any] = encoder_attention_heads UpperCamelCase__: Tuple = decoder_ffn_dim UpperCamelCase__: int = decoder_layers UpperCamelCase__: List[str] = decoder_attention_heads UpperCamelCase__: int = dropout UpperCamelCase__: List[str] = attention_dropout UpperCamelCase__: Tuple = activation_dropout UpperCamelCase__: Optional[int] = activation_function UpperCamelCase__: Dict = init_std UpperCamelCase__: Optional[Any] = encoder_layerdrop UpperCamelCase__: Any = decoder_layerdrop UpperCamelCase__: Optional[int] = use_cache UpperCamelCase__: Optional[int] = encoder_layers UpperCamelCase__: List[Any] = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( pad_token_id=__lowerCamelCase , eos_token_id=__lowerCamelCase , is_encoder_decoder=__lowerCamelCase , decoder_start_token_id=__lowerCamelCase , forced_eos_token_id=__lowerCamelCase , **__lowerCamelCase , ) @property def UpperCAmelCase_ ( self: Any ): '''simple docstring''' return self.encoder_attention_heads @property def UpperCAmelCase_ ( self: Tuple ): '''simple docstring''' return self.d_model
149
def lowerCAmelCase_ ( A_ ,A_): if b == 0: return 1 if (b % 2) == 0: return actual_power(A_ ,int(b / 2)) * actual_power(A_ ,int(b / 2)) else: return a * actual_power(A_ ,int(b / 2)) * actual_power(A_ ,int(b / 2)) def lowerCAmelCase_ ( A_ ,A_): if b < 0: return 1 / actual_power(A_ ,A_) return actual_power(A_ ,A_) if __name__ == "__main__": print(power(-2, -3))
149
1
def snake_case_(_UpperCamelCase , _UpperCamelCase ) -> str: """simple docstring""" return int((input_a, input_a).count(0 ) == 0 ) def snake_case_() -> Any: """simple docstring""" assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_gate(1 , 0 ) == 0 assert and_gate(1 , 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
369
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging __A = logging.get_logger(__name__) __A = '''▁''' __A = { '''vocab_file''': '''vocab.json''', '''spm_file''': '''sentencepiece.bpe.model''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __A = { '''vocab_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json''', }, '''spm_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model''', }, '''tokenizer_config_file''': { '''facebook/m2m100_418M''': '''https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json''', '''facebook/m2m100_1.2B''': '''https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json''', }, } __A = { '''facebook/m2m100_418M''': 10_24, } # fmt: off __A = { '''m2m100''': ['''af''', '''am''', '''ar''', '''ast''', '''az''', '''ba''', '''be''', '''bg''', '''bn''', '''br''', '''bs''', '''ca''', '''ceb''', '''cs''', '''cy''', '''da''', '''de''', '''el''', '''en''', '''es''', '''et''', '''fa''', '''ff''', '''fi''', '''fr''', '''fy''', '''ga''', '''gd''', '''gl''', '''gu''', '''ha''', '''he''', '''hi''', '''hr''', '''ht''', '''hu''', '''hy''', '''id''', '''ig''', '''ilo''', '''is''', '''it''', '''ja''', '''jv''', '''ka''', '''kk''', '''km''', '''kn''', '''ko''', '''lb''', '''lg''', '''ln''', '''lo''', '''lt''', '''lv''', '''mg''', '''mk''', '''ml''', '''mn''', '''mr''', '''ms''', '''my''', '''ne''', '''nl''', '''no''', '''ns''', '''oc''', '''or''', '''pa''', '''pl''', '''ps''', '''pt''', '''ro''', '''ru''', '''sd''', '''si''', '''sk''', '''sl''', '''so''', '''sq''', '''sr''', '''ss''', '''su''', '''sv''', '''sw''', '''ta''', '''th''', '''tl''', '''tn''', '''tr''', '''uk''', '''ur''', '''uz''', '''vi''', '''wo''', '''xh''', '''yi''', '''yo''', '''zh''', '''zu'''], '''wmt21''': ['''en''', '''ha''', '''is''', '''ja''', '''cs''', '''ru''', '''zh''', '''de'''] } class lowercase_ ( __lowercase ): UpperCamelCase_ : str = VOCAB_FILES_NAMES UpperCamelCase_ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ : Dict = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ : Union[str, Any] = ["input_ids", "attention_mask"] UpperCamelCase_ : List[int] = [] UpperCamelCase_ : List[int] = [] def __init__( self : str , A__ : str , A__ : Optional[Any] , A__ : Union[str, Any]=None , A__ : Dict=None , A__ : Any="<s>" , A__ : Union[str, Any]="</s>" , A__ : Tuple="</s>" , A__ : Dict="<pad>" , A__ : List[Any]="<unk>" , A__ : str="m2m100" , A__ : Optional[Dict[str, Any]] = None , A__ : List[Any]=8 , **A__ : Union[str, Any] , ) -> None: _snake_case = {} if sp_model_kwargs is None else sp_model_kwargs _snake_case = language_codes _snake_case = FAIRSEQ_LANGUAGE_CODES[language_codes] _snake_case = {lang_code: f"""__{lang_code}__""" for lang_code in fairseq_language_code} _snake_case = kwargs.get('''additional_special_tokens''' , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(A__ ) for lang_code in fairseq_language_code if self.get_lang_token(A__ ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=A__ , tgt_lang=A__ , bos_token=A__ , eos_token=A__ , sep_token=A__ , unk_token=A__ , pad_token=A__ , language_codes=A__ , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=A__ , **A__ , ) _snake_case = vocab_file _snake_case = load_json(A__ ) _snake_case = {v: k for k, v in self.encoder.items()} _snake_case = spm_file _snake_case = load_spm(A__ , self.sp_model_kwargs ) _snake_case = len(self.encoder ) _snake_case = { self.get_lang_token(A__ ): self.encoder_size + i for i, lang_code in enumerate(A__ ) } _snake_case = {lang_code: self.encoder_size + i for i, lang_code in enumerate(A__ )} _snake_case = {v: k for k, v in self.lang_token_to_id.items()} _snake_case = src_lang if src_lang is not None else '''en''' _snake_case = tgt_lang _snake_case = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) _snake_case = num_madeup_words @property def UpperCamelCase_ ( self : int ) -> int: return len(self.encoder ) + len(self.lang_token_to_id ) @property def UpperCamelCase_ ( self : Dict ) -> str: return self._src_lang @src_lang.setter def UpperCamelCase_ ( self : List[str] , A__ : str ) -> None: _snake_case = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase_ ( self : Any , A__ : str ) -> List[str]: return self.sp_model.encode(A__ , out_type=A__ ) def UpperCamelCase_ ( self : Optional[int] , A__ : Dict ) -> str: if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(A__ , self.encoder[self.unk_token] ) def UpperCamelCase_ ( self : Union[str, Any] , A__ : int ) -> str: if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(A__ , self.unk_token ) def UpperCamelCase_ ( self : Optional[int] , A__ : Optional[int] ) -> List[Any]: _snake_case = [] _snake_case = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(A__ ) + token _snake_case = [] else: current_sub_tokens.append(A__ ) out_string += self.sp_model.decode(A__ ) return out_string.strip() def UpperCamelCase_ ( self : str , A__ : List[int] , A__ : Optional[List[int]] = None , A__ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=A__ , token_ids_a=A__ , already_has_special_tokens=A__ ) _snake_case = [1] * len(self.prefix_tokens ) _snake_case = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(A__ )) + suffix_ones return prefix_ones + ([0] * len(A__ )) + ([0] * len(A__ )) + suffix_ones def UpperCamelCase_ ( self : Tuple , A__ : List[int] , A__ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase_ ( self : str ) -> Dict: _snake_case = {self.convert_ids_to_tokens(A__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ) -> Dict: _snake_case = self.__dict__.copy() _snake_case = None return state def __setstate__( self : Union[str, Any] , A__ : Dict ) -> None: _snake_case = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): _snake_case = {} _snake_case = load_spm(self.spm_file , self.sp_model_kwargs ) def UpperCamelCase_ ( self : Any , A__ : str , A__ : Optional[str] = None ) -> Tuple[str]: _snake_case = Path(A__ ) if not save_dir.is_dir(): raise OSError(f"""{save_directory} should be a directory""" ) _snake_case = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''vocab_file'''] ) _snake_case = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''spm_file'''] ) save_json(self.encoder , A__ ) if os.path.abspath(self.spm_file ) != os.path.abspath(A__ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , A__ ) elif not os.path.isfile(self.spm_file ): with open(A__ , '''wb''' ) as fi: _snake_case = self.sp_model.serialized_model_proto() fi.write(A__ ) return (str(A__ ), str(A__ )) def UpperCamelCase_ ( self : Optional[int] , A__ : List[str] , A__ : str = "en" , A__ : Optional[List[str]] = None , A__ : str = "ro" , **A__ : List[Any] , ) -> BatchEncoding: _snake_case = src_lang _snake_case = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(A__ , A__ , **A__ ) def UpperCamelCase_ ( self : List[str] , A__ : int , A__ : Optional[str] , A__ : Optional[str] , **A__ : Union[str, Any] ) -> Tuple: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) _snake_case = src_lang _snake_case = self(A__ , add_special_tokens=A__ , **A__ ) _snake_case = self.get_lang_id(A__ ) _snake_case = tgt_lang_id return inputs def UpperCamelCase_ ( self : Dict ) -> Optional[Any]: self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase_ ( self : Optional[Any] ) -> Dict: self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase_ ( self : List[Any] , A__ : str ) -> None: _snake_case = self.get_lang_token(A__ ) _snake_case = self.lang_token_to_id[lang_token] _snake_case = [self.cur_lang_id] _snake_case = [self.eos_token_id] def UpperCamelCase_ ( self : List[str] , A__ : str ) -> None: _snake_case = self.get_lang_token(A__ ) _snake_case = self.lang_token_to_id[lang_token] _snake_case = [self.cur_lang_id] _snake_case = [self.eos_token_id] def UpperCamelCase_ ( self : Dict , A__ : str ) -> str: return self.lang_code_to_token[lang] def UpperCamelCase_ ( self : Tuple , A__ : str ) -> int: _snake_case = self.get_lang_token(A__ ) return self.lang_token_to_id[lang_token] def snake_case_(_UpperCamelCase , _UpperCamelCase ) -> sentencepiece.SentencePieceProcessor: """simple docstring""" _snake_case = sentencepiece.SentencePieceProcessor(**_UpperCamelCase ) spm.Load(str(_UpperCamelCase ) ) return spm def snake_case_(_UpperCamelCase ) -> Union[Dict, List]: """simple docstring""" with open(_UpperCamelCase , '''r''' ) as f: return json.load(_UpperCamelCase ) def snake_case_(_UpperCamelCase , _UpperCamelCase ) -> None: """simple docstring""" with open(_UpperCamelCase , '''w''' ) as f: json.dump(_UpperCamelCase , _UpperCamelCase , indent=2 )
278
0
'''simple docstring''' from pathlib import Path import fire def __magic_name__( lowerCamelCase, lowerCamelCase, lowerCamelCase): __lowerCAmelCase = Path(lowerCamelCase) __lowerCAmelCase = Path(lowerCamelCase) dest_dir.mkdir(exist_ok=lowerCamelCase) for path in src_dir.iterdir(): __lowerCAmelCase = [x.rstrip() for x in list(path.open().readlines())][:n] __lowerCAmelCase = dest_dir.joinpath(path.name) print(lowerCamelCase) dest_path.open('''w''').write('''\n'''.join(lowerCamelCase)) if __name__ == "__main__": fire.Fire(minify)
174
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class a__ ( __A ): """simple docstring""" __UpperCamelCase : torch.FloatTensor class a__ ( __A , __A ): """simple docstring""" @register_to_config def __init__(self , __lowercase = 3 , __lowercase = 3 , __lowercase = ("DownEncoderBlock2D",) , __lowercase = ("UpDecoderBlock2D",) , __lowercase = (64,) , __lowercase = 1 , __lowercase = "silu" , __lowercase = 3 , __lowercase = 32 , __lowercase = 2_56 , __lowercase = 32 , __lowercase = None , __lowercase = 0.1_8_2_1_5 , __lowercase = "group" , ): super().__init__() # pass init params to Encoder __lowerCAmelCase = Encoder( in_channels=__lowercase , out_channels=__lowercase , down_block_types=__lowercase , block_out_channels=__lowercase , layers_per_block=__lowercase , act_fn=__lowercase , norm_num_groups=__lowercase , double_z=__lowercase , ) __lowerCAmelCase = vq_embed_dim if vq_embed_dim is not None else latent_channels __lowerCAmelCase = nn.Convad(__lowercase , __lowercase , 1 ) __lowerCAmelCase = VectorQuantizer(__lowercase , __lowercase , beta=0.2_5 , remap=__lowercase , sane_index_shape=__lowercase ) __lowerCAmelCase = nn.Convad(__lowercase , __lowercase , 1 ) # pass init params to Decoder __lowerCAmelCase = Decoder( in_channels=__lowercase , out_channels=__lowercase , up_block_types=__lowercase , block_out_channels=__lowercase , layers_per_block=__lowercase , act_fn=__lowercase , norm_num_groups=__lowercase , norm_type=__lowercase , ) @apply_forward_hook def _snake_case (self , __lowercase , __lowercase = True ): __lowerCAmelCase = self.encoder(__lowercase ) __lowerCAmelCase = self.quant_conv(__lowercase ) if not return_dict: return (h,) return VQEncoderOutput(latents=__lowercase ) @apply_forward_hook def _snake_case (self , __lowercase , __lowercase = False , __lowercase = True ): # also go through quantization layer if not force_not_quantize: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = self.quantize(__lowercase ) else: __lowerCAmelCase = h __lowerCAmelCase = self.post_quant_conv(__lowercase ) __lowerCAmelCase = self.decoder(__lowercase , quant if self.config.norm_type == '''spatial''' else None ) if not return_dict: return (dec,) return DecoderOutput(sample=__lowercase ) def _snake_case (self , __lowercase , __lowercase = True ): __lowerCAmelCase = sample __lowerCAmelCase = self.encode(__lowercase ).latents __lowerCAmelCase = self.decode(__lowercase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=__lowercase )
174
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { """microsoft/table-transformer-detection""": ( """https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json""" ), } class UpperCamelCase__ ( lowerCAmelCase_ ): '''simple docstring''' __snake_case : List[str] = "table-transformer" __snake_case : Any = ["past_key_values"] __snake_case : Optional[Any] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self : List[str] ,lowerCamelCase__ : Optional[int]=True ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : Optional[int]=3 ,lowerCamelCase__ : Optional[int]=100 ,lowerCamelCase__ : Optional[int]=6 ,lowerCamelCase__ : Optional[int]=2048 ,lowerCamelCase__ : List[str]=8 ,lowerCamelCase__ : List[str]=6 ,lowerCamelCase__ : Dict=2048 ,lowerCamelCase__ : int=8 ,lowerCamelCase__ : Dict=0.0 ,lowerCamelCase__ : Optional[Any]=0.0 ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Optional[Any]="relu" ,lowerCamelCase__ : int=256 ,lowerCamelCase__ : List[str]=0.1 ,lowerCamelCase__ : int=0.0 ,lowerCamelCase__ : int=0.0 ,lowerCamelCase__ : List[str]=0.02 ,lowerCamelCase__ : int=1.0 ,lowerCamelCase__ : str=False ,lowerCamelCase__ : Any="sine" ,lowerCamelCase__ : Union[str, Any]="resnet50" ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Union[str, Any]=False ,lowerCamelCase__ : Optional[int]=1 ,lowerCamelCase__ : Tuple=5 ,lowerCamelCase__ : Optional[Any]=2 ,lowerCamelCase__ : int=1 ,lowerCamelCase__ : int=1 ,lowerCamelCase__ : Union[str, Any]=5 ,lowerCamelCase__ : Optional[int]=2 ,lowerCamelCase__ : Optional[int]=0.1 ,**lowerCamelCase__ : List[Any] ,) -> Any: '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) SCREAMING_SNAKE_CASE = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(lowerCamelCase__ ,lowerCamelCase__ ): SCREAMING_SNAKE_CASE = backbone_config.get("""model_type""" ) SCREAMING_SNAKE_CASE = CONFIG_MAPPING[backbone_model_type] SCREAMING_SNAKE_CASE = config_class.from_dict(lowerCamelCase__ ) # set timm attributes to None SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE, SCREAMING_SNAKE_CASE = None, None, None SCREAMING_SNAKE_CASE = use_timm_backbone SCREAMING_SNAKE_CASE = backbone_config SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = num_queries SCREAMING_SNAKE_CASE = d_model SCREAMING_SNAKE_CASE = encoder_ffn_dim SCREAMING_SNAKE_CASE = encoder_layers SCREAMING_SNAKE_CASE = encoder_attention_heads SCREAMING_SNAKE_CASE = decoder_ffn_dim SCREAMING_SNAKE_CASE = decoder_layers SCREAMING_SNAKE_CASE = decoder_attention_heads SCREAMING_SNAKE_CASE = dropout SCREAMING_SNAKE_CASE = attention_dropout SCREAMING_SNAKE_CASE = activation_dropout SCREAMING_SNAKE_CASE = activation_function SCREAMING_SNAKE_CASE = init_std SCREAMING_SNAKE_CASE = init_xavier_std SCREAMING_SNAKE_CASE = encoder_layerdrop SCREAMING_SNAKE_CASE = decoder_layerdrop SCREAMING_SNAKE_CASE = encoder_layers SCREAMING_SNAKE_CASE = auxiliary_loss SCREAMING_SNAKE_CASE = position_embedding_type SCREAMING_SNAKE_CASE = backbone SCREAMING_SNAKE_CASE = use_pretrained_backbone SCREAMING_SNAKE_CASE = dilation # Hungarian matcher SCREAMING_SNAKE_CASE = class_cost SCREAMING_SNAKE_CASE = bbox_cost SCREAMING_SNAKE_CASE = giou_cost # Loss coefficients SCREAMING_SNAKE_CASE = mask_loss_coefficient SCREAMING_SNAKE_CASE = dice_loss_coefficient SCREAMING_SNAKE_CASE = bbox_loss_coefficient SCREAMING_SNAKE_CASE = giou_loss_coefficient SCREAMING_SNAKE_CASE = eos_coefficient super().__init__(is_encoder_decoder=lowerCamelCase__ ,**lowerCamelCase__ ) @property def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> int: '''simple docstring''' return self.encoder_attention_heads @property def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> int: '''simple docstring''' return self.d_model class UpperCamelCase__ ( lowerCAmelCase_ ): '''simple docstring''' __snake_case : int = version.parse("1.11" ) @property def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> float: '''simple docstring''' return 1e-5 @property def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> int: '''simple docstring''' return 12
358
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class UpperCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[Any]=7 ,lowerCamelCase__ : Optional[Any]=3 ,lowerCamelCase__ : Union[str, Any]=18 ,lowerCamelCase__ : Optional[int]=30 ,lowerCamelCase__ : int=400 ,lowerCamelCase__ : List[str]=True ,lowerCamelCase__ : Union[str, Any]=None ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Optional[int]=[0.5, 0.5, 0.5] ,lowerCamelCase__ : Any=[0.5, 0.5, 0.5] ,) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = size if size is not None else {"""height""": 18, """width""": 18} SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = image_size SCREAMING_SNAKE_CASE = min_resolution SCREAMING_SNAKE_CASE = max_resolution SCREAMING_SNAKE_CASE = do_resize SCREAMING_SNAKE_CASE = size SCREAMING_SNAKE_CASE = do_normalize SCREAMING_SNAKE_CASE = image_mean SCREAMING_SNAKE_CASE = image_std def SCREAMING_SNAKE_CASE__ ( self : Any ) -> Union[str, Any]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class UpperCamelCase__ ( lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' __snake_case : Tuple = DPTImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = DPTImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> int: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ ,"""image_mean""" ) ) self.assertTrue(hasattr(lowerCamelCase__ ,"""image_std""" ) ) self.assertTrue(hasattr(lowerCamelCase__ ,"""do_normalize""" ) ) self.assertTrue(hasattr(lowerCamelCase__ ,"""do_resize""" ) ) self.assertTrue(hasattr(lowerCamelCase__ ,"""size""" ) ) def SCREAMING_SNAKE_CASE__ ( self : int ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{"""height""": 18, """width""": 18} ) SCREAMING_SNAKE_CASE = self.image_processing_class.from_dict(self.image_processor_dict ,size=42 ) self.assertEqual(image_processor.size ,{"""height""": 42, """width""": 42} ) def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ ,Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched SCREAMING_SNAKE_CASE = image_processing(lowerCamelCase__ ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def SCREAMING_SNAKE_CASE__ ( self : List[Any] ) -> int: '''simple docstring''' SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowerCamelCase__ ,numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ ,np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched SCREAMING_SNAKE_CASE = image_processing(lowerCamelCase__ ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowerCamelCase__ ,torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ ,torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE = image_processing(image_inputs[0] ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,) # Test batched SCREAMING_SNAKE_CASE = image_processing(lowerCamelCase__ ,return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["""height"""], self.image_processor_tester.size["""width"""], ) ,)
193
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class snake_case_( unittest.TestCase ): def lowerCamelCase__ ( self : List[Any] ): lowerCAmelCase : Optional[int] = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 1_2_8, '''min_length''': 1_2, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 1_4_2, '''min_length''': 5_6, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 6_2, '''min_length''': 1_1, '''num_beams''': 6}, } } lowerCAmelCase : int = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 1_2_8, '''task_specific_params.summarization.min_length''': 1_2, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 1_4_2, '''task_specific_params.summarization_cnn.min_length''': 5_6, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 6_2, '''task_specific_params.summarization_xsum.min_length''': 1_1, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(UpperCamelCase_ ) , UpperCamelCase_ ) def lowerCamelCase__ ( self : Dict ): lowerCAmelCase : Dict = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ ) , x.transpose() ) ) lowerCAmelCase : List[Any] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def lowerCamelCase__ ( self : Optional[int] ): lowerCAmelCase : Optional[int] = np.random.randn(3 , 4 ) lowerCAmelCase : Any = torch.tensor(UpperCamelCase_ ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ ) , transpose(UpperCamelCase_ ).numpy() ) ) lowerCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) lowerCAmelCase : Tuple = torch.tensor(UpperCamelCase_ ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ , axes=(1, 2, 0) ) , transpose(UpperCamelCase_ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def lowerCamelCase__ ( self : Tuple ): lowerCAmelCase : Any = np.random.randn(3 , 4 ) lowerCAmelCase : str = tf.constant(UpperCamelCase_ ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ ) , transpose(UpperCamelCase_ ).numpy() ) ) lowerCAmelCase : Optional[Any] = np.random.randn(3 , 4 , 5 ) lowerCAmelCase : List[Any] = tf.constant(UpperCamelCase_ ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ , axes=(1, 2, 0) ) , transpose(UpperCamelCase_ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def lowerCamelCase__ ( self : List[str] ): lowerCAmelCase : Optional[Any] = np.random.randn(3 , 4 ) lowerCAmelCase : Union[str, Any] = jnp.array(UpperCamelCase_ ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ ) , np.asarray(transpose(UpperCamelCase_ ) ) ) ) lowerCAmelCase : Tuple = np.random.randn(3 , 4 , 5 ) lowerCAmelCase : Dict = jnp.array(UpperCamelCase_ ) self.assertTrue(np.allclose(transpose(UpperCamelCase_ , axes=(1, 2, 0) ) , np.asarray(transpose(UpperCamelCase_ , axes=(1, 2, 0) ) ) ) ) def lowerCamelCase__ ( self : str ): lowerCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (4, 3) ) , np.reshape(UpperCamelCase_ , (4, 3) ) ) ) lowerCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (1_2, 5) ) , np.reshape(UpperCamelCase_ , (1_2, 5) ) ) ) @require_torch def lowerCamelCase__ ( self : Any ): lowerCAmelCase : Optional[Any] = np.random.randn(3 , 4 ) lowerCAmelCase : Optional[Any] = torch.tensor(UpperCamelCase_ ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (4, 3) ) , reshape(UpperCamelCase_ , (4, 3) ).numpy() ) ) lowerCAmelCase : List[Any] = np.random.randn(3 , 4 , 5 ) lowerCAmelCase : Union[str, Any] = torch.tensor(UpperCamelCase_ ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (1_2, 5) ) , reshape(UpperCamelCase_ , (1_2, 5) ).numpy() ) ) @require_tf def lowerCamelCase__ ( self : Optional[int] ): lowerCAmelCase : List[Any] = np.random.randn(3 , 4 ) lowerCAmelCase : str = tf.constant(UpperCamelCase_ ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (4, 3) ) , reshape(UpperCamelCase_ , (4, 3) ).numpy() ) ) lowerCAmelCase : int = np.random.randn(3 , 4 , 5 ) lowerCAmelCase : Tuple = tf.constant(UpperCamelCase_ ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (1_2, 5) ) , reshape(UpperCamelCase_ , (1_2, 5) ).numpy() ) ) @require_flax def lowerCamelCase__ ( self : Optional[Any] ): lowerCAmelCase : Dict = np.random.randn(3 , 4 ) lowerCAmelCase : Union[str, Any] = jnp.array(UpperCamelCase_ ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (4, 3) ) , np.asarray(reshape(UpperCamelCase_ , (4, 3) ) ) ) ) lowerCAmelCase : List[str] = np.random.randn(3 , 4 , 5 ) lowerCAmelCase : int = jnp.array(UpperCamelCase_ ) self.assertTrue(np.allclose(reshape(UpperCamelCase_ , (1_2, 5) ) , np.asarray(reshape(UpperCamelCase_ , (1_2, 5) ) ) ) ) def lowerCamelCase__ ( self : List[str] ): lowerCAmelCase : List[Any] = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ ) , np.squeeze(UpperCamelCase_ ) ) ) lowerCAmelCase : Any = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ , axis=2 ) , np.squeeze(UpperCamelCase_ , axis=2 ) ) ) @require_torch def lowerCamelCase__ ( self : int ): lowerCAmelCase : Any = np.random.randn(1 , 3 , 4 ) lowerCAmelCase : Union[str, Any] = torch.tensor(UpperCamelCase_ ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ ) , squeeze(UpperCamelCase_ ).numpy() ) ) lowerCAmelCase : Dict = np.random.randn(1 , 4 , 1 , 5 ) lowerCAmelCase : Tuple = torch.tensor(UpperCamelCase_ ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ , axis=2 ) , squeeze(UpperCamelCase_ , axis=2 ).numpy() ) ) @require_tf def lowerCamelCase__ ( self : Optional[Any] ): lowerCAmelCase : str = np.random.randn(1 , 3 , 4 ) lowerCAmelCase : List[str] = tf.constant(UpperCamelCase_ ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ ) , squeeze(UpperCamelCase_ ).numpy() ) ) lowerCAmelCase : Optional[Any] = np.random.randn(1 , 4 , 1 , 5 ) lowerCAmelCase : Optional[int] = tf.constant(UpperCamelCase_ ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ , axis=2 ) , squeeze(UpperCamelCase_ , axis=2 ).numpy() ) ) @require_flax def lowerCamelCase__ ( self : Any ): lowerCAmelCase : int = np.random.randn(1 , 3 , 4 ) lowerCAmelCase : str = jnp.array(UpperCamelCase_ ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ ) , np.asarray(squeeze(UpperCamelCase_ ) ) ) ) lowerCAmelCase : List[str] = np.random.randn(1 , 4 , 1 , 5 ) lowerCAmelCase : str = jnp.array(UpperCamelCase_ ) self.assertTrue(np.allclose(squeeze(UpperCamelCase_ , axis=2 ) , np.asarray(squeeze(UpperCamelCase_ , axis=2 ) ) ) ) def lowerCamelCase__ ( self : Tuple ): lowerCAmelCase : List[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(UpperCamelCase_ , axis=1 ) , np.expand_dims(UpperCamelCase_ , axis=1 ) ) ) @require_torch def lowerCamelCase__ ( self : Dict ): lowerCAmelCase : Any = np.random.randn(3 , 4 ) lowerCAmelCase : Any = torch.tensor(UpperCamelCase_ ) self.assertTrue(np.allclose(expand_dims(UpperCamelCase_ , axis=1 ) , expand_dims(UpperCamelCase_ , axis=1 ).numpy() ) ) @require_tf def lowerCamelCase__ ( self : Union[str, Any] ): lowerCAmelCase : Tuple = np.random.randn(3 , 4 ) lowerCAmelCase : Optional[int] = tf.constant(UpperCamelCase_ ) self.assertTrue(np.allclose(expand_dims(UpperCamelCase_ , axis=1 ) , expand_dims(UpperCamelCase_ , axis=1 ).numpy() ) ) @require_flax def lowerCamelCase__ ( self : Optional[int] ): lowerCAmelCase : Dict = np.random.randn(3 , 4 ) lowerCAmelCase : Tuple = jnp.array(UpperCamelCase_ ) self.assertTrue(np.allclose(expand_dims(UpperCamelCase_ , axis=1 ) , np.asarray(expand_dims(UpperCamelCase_ , axis=1 ) ) ) )
60
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _UpperCAmelCase : Union[str, Any] = logging.get_logger(__name__) def UpperCAmelCase__ ( lowerCamelCase ): lowercase :int = DPTConfig(embedding_type="hybrid" ) if "large" in checkpoint_url: lowercase :Tuple = 1024 lowercase :Tuple = 4096 lowercase :Union[str, Any] = 24 lowercase :Tuple = 16 lowercase :Optional[Any] = [5, 11, 17, 23] lowercase :Optional[Any] = [256, 512, 1024, 1024] lowercase :Tuple = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: lowercase :Any = 768 lowercase :Optional[int] = [1, 1, 1, 0.5] lowercase :int = [256, 512, 768, 768] lowercase :str = 150 lowercase :Tuple = 16 lowercase :List[str] = (1, 384, 384) lowercase :Any = False lowercase :Optional[int] = "project" if "ade" in checkpoint_url: lowercase :Dict = True lowercase :str = 768 lowercase :Tuple = [1, 1, 1, 0.5] lowercase :Dict = 150 lowercase :Any = 16 lowercase :List[Any] = "huggingface/label-files" lowercase :Union[str, Any] = "ade20k-id2label.json" lowercase :List[str] = json.load(open(cached_download(hf_hub_url(lowerCamelCase, lowerCamelCase, repo_type="dataset" ) ), "r" ) ) lowercase :Optional[Any] = {int(lowerCamelCase ): v for k, v in idalabel.items()} lowercase :Union[str, Any] = idalabel lowercase :Optional[Any] = {v: k for k, v in idalabel.items()} lowercase :List[str] = [1, 150, 480, 480] return config, expected_shape def UpperCAmelCase__ ( lowerCamelCase ): lowercase :List[str] = ["pretrained.model.head.weight", "pretrained.model.head.bias"] for k in ignore_keys: state_dict.pop(lowerCamelCase, lowerCamelCase ) def UpperCAmelCase__ ( lowerCamelCase ): if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): lowercase :Any = name.replace("pretrained.model", "dpt.encoder" ) if "pretrained.model" in name: lowercase :int = name.replace("pretrained.model", "dpt.embeddings" ) if "patch_embed" in name: lowercase :Any = name.replace("patch_embed", "" ) if "pos_embed" in name: lowercase :Optional[Any] = name.replace("pos_embed", "position_embeddings" ) if "attn.proj" in name: lowercase :List[str] = name.replace("attn.proj", "attention.output.dense" ) if "proj" in name and "project" not in name: lowercase :str = name.replace("proj", "projection" ) if "blocks" in name: lowercase :List[Any] = name.replace("blocks", "layer" ) if "mlp.fc1" in name: lowercase :str = name.replace("mlp.fc1", "intermediate.dense" ) if "mlp.fc2" in name: lowercase :Any = name.replace("mlp.fc2", "output.dense" ) if "norm1" in name and "backbone" not in name: lowercase :Any = name.replace("norm1", "layernorm_before" ) if "norm2" in name and "backbone" not in name: lowercase :Any = name.replace("norm2", "layernorm_after" ) if "scratch.output_conv" in name: lowercase :Optional[Any] = name.replace("scratch.output_conv", "head" ) if "scratch" in name: lowercase :Union[str, Any] = name.replace("scratch", "neck" ) if "layer1_rn" in name: lowercase :str = name.replace("layer1_rn", "convs.0" ) if "layer2_rn" in name: lowercase :Dict = name.replace("layer2_rn", "convs.1" ) if "layer3_rn" in name: lowercase :Tuple = name.replace("layer3_rn", "convs.2" ) if "layer4_rn" in name: lowercase :Optional[int] = name.replace("layer4_rn", "convs.3" ) if "refinenet" in name: lowercase :Any = int(name[len("neck.refinenet" ) : len("neck.refinenet" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 lowercase :Any = name.replace(F"refinenet{layer_idx}", F"fusion_stage.layers.{abs(layer_idx-4 )}" ) if "out_conv" in name: lowercase :str = name.replace("out_conv", "projection" ) if "resConfUnit1" in name: lowercase :Optional[Any] = name.replace("resConfUnit1", "residual_layer1" ) if "resConfUnit2" in name: lowercase :int = name.replace("resConfUnit2", "residual_layer2" ) if "conv1" in name: lowercase :Optional[int] = name.replace("conv1", "convolution1" ) if "conv2" in name: lowercase :List[str] = name.replace("conv2", "convolution2" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: lowercase :Tuple = name.replace("pretrained.act_postprocess1.0.project.0", "neck.reassemble_stage.readout_projects.0.0" ) if "pretrained.act_postprocess2.0.project.0" in name: lowercase :Dict = name.replace("pretrained.act_postprocess2.0.project.0", "neck.reassemble_stage.readout_projects.1.0" ) if "pretrained.act_postprocess3.0.project.0" in name: lowercase :Any = name.replace("pretrained.act_postprocess3.0.project.0", "neck.reassemble_stage.readout_projects.2.0" ) if "pretrained.act_postprocess4.0.project.0" in name: lowercase :Optional[Any] = name.replace("pretrained.act_postprocess4.0.project.0", "neck.reassemble_stage.readout_projects.3.0" ) # resize blocks if "pretrained.act_postprocess1.3" in name: lowercase :Dict = name.replace("pretrained.act_postprocess1.3", "neck.reassemble_stage.layers.0.projection" ) if "pretrained.act_postprocess1.4" in name: lowercase :Any = name.replace("pretrained.act_postprocess1.4", "neck.reassemble_stage.layers.0.resize" ) if "pretrained.act_postprocess2.3" in name: lowercase :List[Any] = name.replace("pretrained.act_postprocess2.3", "neck.reassemble_stage.layers.1.projection" ) if "pretrained.act_postprocess2.4" in name: lowercase :str = name.replace("pretrained.act_postprocess2.4", "neck.reassemble_stage.layers.1.resize" ) if "pretrained.act_postprocess3.3" in name: lowercase :str = name.replace("pretrained.act_postprocess3.3", "neck.reassemble_stage.layers.2.projection" ) if "pretrained.act_postprocess4.3" in name: lowercase :List[Any] = name.replace("pretrained.act_postprocess4.3", "neck.reassemble_stage.layers.3.projection" ) if "pretrained.act_postprocess4.4" in name: lowercase :Any = name.replace("pretrained.act_postprocess4.4", "neck.reassemble_stage.layers.3.resize" ) if "pretrained" in name: lowercase :Any = name.replace("pretrained", "dpt" ) if "bn" in name: lowercase :Optional[int] = name.replace("bn", "batch_norm" ) if "head" in name: lowercase :Union[str, Any] = name.replace("head", "head.head" ) if "encoder.norm" in name: lowercase :Optional[Any] = name.replace("encoder.norm", "layernorm" ) if "auxlayer" in name: lowercase :str = name.replace("auxlayer", "auxiliary_head.head" ) if "backbone" in name: lowercase :List[str] = name.replace("backbone", "backbone.bit.encoder" ) if ".." in name: lowercase :Optional[int] = name.replace("..", "." ) if "stem.conv" in name: lowercase :Union[str, Any] = name.replace("stem.conv", "bit.embedder.convolution" ) if "blocks" in name: lowercase :List[str] = name.replace("blocks", "layers" ) if "convolution" in name and "backbone" in name: lowercase :int = name.replace("convolution", "conv" ) if "layer" in name and "backbone" in name: lowercase :List[str] = name.replace("layer", "layers" ) if "backbone.bit.encoder.bit" in name: lowercase :Dict = name.replace("backbone.bit.encoder.bit", "backbone.bit" ) if "embedder.conv" in name: lowercase :Dict = name.replace("embedder.conv", "embedder.convolution" ) if "backbone.bit.encoder.stem.norm" in name: lowercase :str = name.replace("backbone.bit.encoder.stem.norm", "backbone.bit.embedder.norm" ) return name def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase ): for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowercase :str = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.weight" ) lowercase :str = state_dict.pop(F"dpt.encoder.layer.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict lowercase :List[str] = in_proj_weight[: config.hidden_size, :] lowercase :int = in_proj_bias[: config.hidden_size] lowercase :Union[str, Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowercase :Union[str, Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowercase :Dict = in_proj_weight[ -config.hidden_size :, : ] lowercase :List[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase__ ( ): lowercase :List[str] = "http://images.cocodataset.org/val2017/000000039769.jpg" lowercase :Any = Image.open(requests.get(lowerCamelCase, stream=lowerCamelCase ).raw ) return im @torch.no_grad() def UpperCAmelCase__ ( lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase, lowerCamelCase ): lowercase , lowercase :Optional[int] = get_dpt_config(lowerCamelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") lowercase :List[str] = torch.load(lowerCamelCase, map_location="cpu" ) # remove certain keys remove_ignore_keys_(lowerCamelCase ) # rename keys for key in state_dict.copy().keys(): lowercase :int = state_dict.pop(lowerCamelCase ) lowercase :Dict = val # read in qkv matrices read_in_q_k_v(lowerCamelCase, lowerCamelCase ) # load HuggingFace model lowercase :int = DPTForSemanticSegmentation(lowerCamelCase ) if "ade" in checkpoint_url else DPTForDepthEstimation(lowerCamelCase ) model.load_state_dict(lowerCamelCase ) model.eval() # Check outputs on an image lowercase :Union[str, Any] = 480 if "ade" in checkpoint_url else 384 lowercase :Optional[int] = DPTImageProcessor(size=lowerCamelCase ) lowercase :Dict = prepare_img() lowercase :Optional[Any] = image_processor(lowerCamelCase, return_tensors="pt" ) # forward pass lowercase :Any = model(**lowerCamelCase ).logits if "ade" in checkpoint_url else model(**lowerCamelCase ).predicted_depth if show_prediction: lowercase :List[Any] = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ), size=(image.size[1], image.size[0]), mode="bicubic", align_corners=lowerCamelCase, ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(lowerCamelCase ).mkdir(exist_ok=lowerCamelCase ) print(F"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(lowerCamelCase ) print(F"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(lowerCamelCase ) if push_to_hub: model.push_to_hub("ybelkada/dpt-hybrid-midas" ) image_processor.push_to_hub("ybelkada/dpt-hybrid-midas" ) if __name__ == "__main__": _UpperCAmelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt", type=str, help="URL of the original DPT checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", action="store_true", ) parser.add_argument( "--model_name", default="dpt-large", type=str, help="Name of the model, in case you're pushing to the hub.", ) parser.add_argument( "--show_prediction", action="store_true", ) _UpperCAmelCase : List[Any] = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
236
0
import os from collections.abc import Iterator def __lowercase ( a__ = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(a__ ): __SCREAMING_SNAKE_CASE = [d for d in dir_names if d != 'scripts' and d[0] not in '._'] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(a__ )[1] in (".py", ".ipynb"): yield os.path.join(a__ , a__ ).lstrip('./' ) def __lowercase ( a__ ) -> int: return f"""{i * " "}*""" if i else "\n##" def __lowercase ( a__ , a__ ) -> str: __SCREAMING_SNAKE_CASE = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(a__ ) or old_parts[i] != new_part) and new_part: print(f"""{md_prefix(a__ )} {new_part.replace("_" , " " ).title()}""" ) return new_path def __lowercase ( a__ = "." ) -> None: __SCREAMING_SNAKE_CASE = '' for filepath in sorted(good_file_paths(a__ ) ): __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = os.path.split(a__ ) if filepath != old_path: __SCREAMING_SNAKE_CASE = print_path(a__ , a__ ) __SCREAMING_SNAKE_CASE = (filepath.count(os.sep ) + 1) if filepath else 0 __SCREAMING_SNAKE_CASE = f"""{filepath}/{filename}""".replace(' ' , '%20' ) __SCREAMING_SNAKE_CASE = os.path.splitext(filename.replace('_' , ' ' ).title() )[0] print(f"""{md_prefix(a__ )} [{filename}]({url})""" ) if __name__ == "__main__": print_directory_md('''.''')
118
def __lowercase ( a__ = 10_00 ) -> int: __SCREAMING_SNAKE_CASE = -1 __SCREAMING_SNAKE_CASE = 0 for a in range(1 , n // 3 ): # Solving the two equations a**2+b**2=c**2 and a+b+c=N eliminating c __SCREAMING_SNAKE_CASE = (n * n - 2 * a * n) // (2 * n - 2 * a) __SCREAMING_SNAKE_CASE = n - a - b if c * c == (a * a + b * b): __SCREAMING_SNAKE_CASE = a * b * c if candidate >= product: __SCREAMING_SNAKE_CASE = candidate return product if __name__ == "__main__": print(F'''{solution() = }''')
118
1
from math import pi, sqrt, tan def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if length < 0 or breadth < 0 or height < 0: raise ValueError('surface_area_cuboid() only accepts non-negative values' ) return 2 * ((length * breadth) + (breadth * height) + (length * height)) def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError('surface_area_sphere() only accepts non-negative values' ) return 4 * pi * radius**2 def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError('surface_area_hemisphere() only accepts non-negative values' ) return 3 * pi * radius**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0 or height < 0: raise ValueError('surface_area_cone() only accepts non-negative values' ) return pi * radius * (radius + (height**2 + radius**2) ** 0.5) def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius_a < 0 or radius_a < 0 or height < 0: raise ValueError( 'surface_area_conical_frustum() only accepts non-negative values' ) __lowerCamelCase = (height**2 + (radius_a - radius_a) ** 2) ** 0.5 return pi * ((slant_height * (radius_a + radius_a)) + radius_a**2 + radius_a**2) def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0 or height < 0: raise ValueError('surface_area_cylinder() only accepts non-negative values' ) return 2 * pi * radius * (height + radius) def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if torus_radius < 0 or tube_radius < 0: raise ValueError('surface_area_torus() only accepts non-negative values' ) if torus_radius < tube_radius: raise ValueError( 'surface_area_torus() does not support spindle or self intersecting tori' ) return 4 * pow(UpperCamelCase__ , 2 ) * torus_radius * tube_radius def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if length < 0 or width < 0: raise ValueError('area_rectangle() only accepts non-negative values' ) return length * width def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if side_length < 0: raise ValueError('area_square() only accepts non-negative values' ) return side_length**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if base < 0 or height < 0: raise ValueError('area_triangle() only accepts non-negative values' ) return (base * height) / 2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if sidea < 0 or sidea < 0 or sidea < 0: raise ValueError('area_triangle_three_sides() only accepts non-negative values' ) elif sidea + sidea < sidea or sidea + sidea < sidea or sidea + sidea < sidea: raise ValueError('Given three sides do not form a triangle' ) __lowerCamelCase = (sidea + sidea + sidea) / 2 __lowerCamelCase = sqrt( semi_perimeter * (semi_perimeter - sidea) * (semi_perimeter - sidea) * (semi_perimeter - sidea) ) return area def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if base < 0 or height < 0: raise ValueError('area_parallelogram() only accepts non-negative values' ) return base * height def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if basea < 0 or basea < 0 or height < 0: raise ValueError('area_trapezium() only accepts non-negative values' ) return 1 / 2 * (basea + basea) * height def lowerCamelCase_ ( UpperCamelCase__ : float ) -> float: """simple docstring""" if radius < 0: raise ValueError('area_circle() only accepts non-negative values' ) return pi * radius**2 def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if radius_x < 0 or radius_y < 0: raise ValueError('area_ellipse() only accepts non-negative values' ) return pi * radius_x * radius_y def lowerCamelCase_ ( UpperCamelCase__ : float , UpperCamelCase__ : float ) -> float: """simple docstring""" if diagonal_a < 0 or diagonal_a < 0: raise ValueError('area_rhombus() only accepts non-negative values' ) return 1 / 2 * diagonal_a * diagonal_a def lowerCamelCase_ ( UpperCamelCase__ : int , UpperCamelCase__ : float ) -> float: """simple docstring""" if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or sides < 3: raise ValueError( 'area_reg_polygon() only accepts integers greater than or \ equal to three as number of sides' ) elif length < 0: raise ValueError( 'area_reg_polygon() only accepts non-negative values as \ length of a side' ) return (sides * length**2) / (4 * tan(pi / sides )) return (sides * length**2) / (4 * tan(pi / sides )) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) # verbose so we can see methods missing tests print("[DEMO] Areas of various geometric shapes: \n") print(f'''Rectangle: {area_rectangle(10, 20) = }''') print(f'''Square: {area_square(10) = }''') print(f'''Triangle: {area_triangle(10, 10) = }''') print(f'''Triangle: {area_triangle_three_sides(5, 12, 13) = }''') print(f'''Parallelogram: {area_parallelogram(10, 20) = }''') print(f'''Rhombus: {area_rhombus(10, 20) = }''') print(f'''Trapezium: {area_trapezium(10, 20, 30) = }''') print(f'''Circle: {area_circle(20) = }''') print(f'''Ellipse: {area_ellipse(10, 20) = }''') print("\nSurface Areas of various geometric shapes: \n") print(f'''Cube: {surface_area_cube(20) = }''') print(f'''Cuboid: {surface_area_cuboid(10, 20, 30) = }''') print(f'''Sphere: {surface_area_sphere(20) = }''') print(f'''Hemisphere: {surface_area_hemisphere(20) = }''') print(f'''Cone: {surface_area_cone(10, 20) = }''') print(f'''Conical Frustum: {surface_area_conical_frustum(10, 20, 30) = }''') print(f'''Cylinder: {surface_area_cylinder(10, 20) = }''') print(f'''Torus: {surface_area_torus(20, 10) = }''') print(f'''Equilateral Triangle: {area_reg_polygon(3, 10) = }''') print(f'''Square: {area_reg_polygon(4, 10) = }''') print(f'''Reqular Pentagon: {area_reg_polygon(5, 10) = }''')
90
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class _SCREAMING_SNAKE_CASE( A ): SCREAMING_SNAKE_CASE_ : torch.FloatTensor SCREAMING_SNAKE_CASE_ : torch.FloatTensor SCREAMING_SNAKE_CASE_ : Optional[torch.FloatTensor] = None class _SCREAMING_SNAKE_CASE( A , A ): SCREAMING_SNAKE_CASE_ : Any = 2 @register_to_config def __init__( self ,SCREAMING_SNAKE_CASE__ = 0.0_2 ,SCREAMING_SNAKE_CASE__ = 1_00 ,SCREAMING_SNAKE_CASE__ = 1.0_0_7 ,SCREAMING_SNAKE_CASE__ = 80 ,SCREAMING_SNAKE_CASE__ = 0.0_5 ,SCREAMING_SNAKE_CASE__ = 50 ,) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Any = sigma_max # setable values __SCREAMING_SNAKE_CASE :int = None __SCREAMING_SNAKE_CASE :np.IntTensor = None __SCREAMING_SNAKE_CASE :torch.FloatTensor = None # sigma(t_i) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ = None ) -> torch.FloatTensor: """simple docstring""" return sample def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ = None ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[Any] = num_inference_steps __SCREAMING_SNAKE_CASE :int = np.arange(0 ,self.num_inference_steps )[::-1].copy() __SCREAMING_SNAKE_CASE :Optional[int] = torch.from_numpy(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Any = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] __SCREAMING_SNAKE_CASE :List[str] = torch.tensor(SCREAMING_SNAKE_CASE__ ,dtype=torch.floataa ,device=SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ = None ) -> Tuple[torch.FloatTensor, float]: """simple docstring""" if self.config.s_min <= sigma <= self.config.s_max: __SCREAMING_SNAKE_CASE :List[str] = min(self.config.s_churn / self.num_inference_steps ,2**0.5 - 1 ) else: __SCREAMING_SNAKE_CASE :Optional[Any] = 0 # sample eps ~ N(0, S_noise^2 * I) __SCREAMING_SNAKE_CASE :Optional[int] = self.config.s_noise * randn_tensor(sample.shape ,generator=SCREAMING_SNAKE_CASE__ ).to(sample.device ) __SCREAMING_SNAKE_CASE :List[str] = sigma + gamma * sigma __SCREAMING_SNAKE_CASE :str = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ = True ,) -> Union[KarrasVeOutput, Tuple]: """simple docstring""" __SCREAMING_SNAKE_CASE :str = sample_hat + sigma_hat * model_output __SCREAMING_SNAKE_CASE :Tuple = (sample_hat - pred_original_sample) / sigma_hat __SCREAMING_SNAKE_CASE :List[Any] = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=SCREAMING_SNAKE_CASE__ ,derivative=SCREAMING_SNAKE_CASE__ ,pred_original_sample=SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ = True ,) -> Union[KarrasVeOutput, Tuple]: """simple docstring""" __SCREAMING_SNAKE_CASE :Any = sample_prev + sigma_prev * model_output __SCREAMING_SNAKE_CASE :List[Any] = (sample_prev - pred_original_sample) / sigma_prev __SCREAMING_SNAKE_CASE :Dict = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=SCREAMING_SNAKE_CASE__ ,derivative=SCREAMING_SNAKE_CASE__ ,pred_original_sample=SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: """simple docstring""" raise NotImplementedError()
191
0
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_torch, require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_torch_available, is_vision_available if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MgpstrProcessor, ViTImageProcessor @require_torch @require_vision class _lowerCamelCase ( unittest.TestCase ): '''simple docstring''' A_ : List[str] = ViTImageProcessor if is_vision_available() else None @property def __lowerCAmelCase ( self : Optional[Any] ) -> List[str]: return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self : int ) -> Dict: __magic_name__ : str = (3, 32, 128) __magic_name__ : Tuple = tempfile.mkdtemp() # fmt: off __magic_name__ : Dict = ['[GO]', '[s]', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z'] # fmt: on __magic_name__ : Union[str, Any] = dict(zip(_A , range(len(_A ) ) ) ) __magic_name__ : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(_A ) + '\n' ) __magic_name__ : Tuple = { 'do_normalize': False, 'do_resize': True, 'image_processor_type': 'ViTImageProcessor', 'resample': 3, 'size': {'height': 32, 'width': 128}, } __magic_name__ : List[str] = os.path.join(self.tmpdirname , _A ) with open(self.image_processor_file , 'w' , encoding='utf-8' ) as fp: json.dump(_A , _A ) def __lowerCAmelCase ( self : Any , **_A : List[str] ) -> Tuple: return MgpstrTokenizer.from_pretrained(self.tmpdirname , **_A ) def __lowerCAmelCase ( self : List[Any] , **_A : Optional[int] ) -> Union[str, Any]: return ViTImageProcessor.from_pretrained(self.tmpdirname , **_A ) def __lowerCAmelCase ( self : List[Any] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self : List[Any] ) -> int: __magic_name__ : Dict = np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta ) __magic_name__ : Union[str, Any] = Image.fromarray(np.moveaxis(_A , 0 , -1 ) ) return image_input def __lowerCAmelCase ( self : str ) -> Optional[Any]: __magic_name__ : Optional[int] = self.get_tokenizer() __magic_name__ : Dict = self.get_image_processor() __magic_name__ : Optional[int] = MgpstrProcessor(tokenizer=_A , image_processor=_A ) processor.save_pretrained(self.tmpdirname ) __magic_name__ : Tuple = MgpstrProcessor.from_pretrained(self.tmpdirname , use_fast=_A ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , _A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor.image_processor , _A ) def __lowerCAmelCase ( self : str ) -> Union[str, Any]: __magic_name__ : Optional[Any] = self.get_tokenizer() __magic_name__ : List[Any] = self.get_image_processor() __magic_name__ : Dict = MgpstrProcessor(tokenizer=_A , image_processor=_A ) processor.save_pretrained(self.tmpdirname ) __magic_name__ : Tuple = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) __magic_name__ : Optional[int] = self.get_image_processor(do_normalize=_A , padding_value=1.0 ) __magic_name__ : List[str] = MgpstrProcessor.from_pretrained( self.tmpdirname , bos_token='(BOS)' , eos_token='(EOS)' , do_normalize=_A , padding_value=1.0 ) self.assertEqual(processor.char_tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.char_tokenizer , _A ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , _A ) def __lowerCAmelCase ( self : Optional[int] ) -> Optional[int]: __magic_name__ : Optional[Any] = self.get_image_processor() __magic_name__ : Optional[Any] = self.get_tokenizer() __magic_name__ : Optional[int] = MgpstrProcessor(tokenizer=_A , image_processor=_A ) __magic_name__ : Dict = self.prepare_image_inputs() __magic_name__ : List[str] = image_processor(_A , return_tensors='np' ) __magic_name__ : Dict = processor(images=_A , return_tensors='np' ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __lowerCAmelCase ( self : Optional[Any] ) -> int: __magic_name__ : Optional[int] = self.get_image_processor() __magic_name__ : Union[str, Any] = self.get_tokenizer() __magic_name__ : Union[str, Any] = MgpstrProcessor(tokenizer=_A , image_processor=_A ) __magic_name__ : Optional[Any] = 'test' __magic_name__ : str = processor(text=_A ) __magic_name__ : List[Any] = tokenizer(_A ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __lowerCAmelCase ( self : Tuple ) -> Dict: __magic_name__ : Dict = self.get_image_processor() __magic_name__ : Optional[Any] = self.get_tokenizer() __magic_name__ : Union[str, Any] = MgpstrProcessor(tokenizer=_A , image_processor=_A ) __magic_name__ : str = 'test' __magic_name__ : int = self.prepare_image_inputs() __magic_name__ : int = processor(text=_A , images=_A ) self.assertListEqual(list(inputs.keys() ) , ['pixel_values', 'labels'] ) # test if it raises when no input is passed with pytest.raises(_A ): processor() def __lowerCAmelCase ( self : Any ) -> Optional[int]: __magic_name__ : Optional[int] = self.get_image_processor() __magic_name__ : Union[str, Any] = self.get_tokenizer() __magic_name__ : Tuple = MgpstrProcessor(tokenizer=_A , image_processor=_A ) __magic_name__ : List[str] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9], [3, 4, 3, 1, 1, 8, 9]] __magic_name__ : Optional[int] = processor.char_decode(_A ) __magic_name__ : Tuple = tokenizer.batch_decode(_A ) __magic_name__ : List[str] = [seq.replace(' ' , '' ) for seq in decoded_tok] self.assertListEqual(_A , _A ) def __lowerCAmelCase ( self : Any ) -> Union[str, Any]: __magic_name__ : int = self.get_image_processor() __magic_name__ : str = self.get_tokenizer() __magic_name__ : Tuple = MgpstrProcessor(tokenizer=_A , image_processor=_A ) __magic_name__ : Tuple = None __magic_name__ : Any = self.prepare_image_inputs() __magic_name__ : Optional[Any] = processor(text=_A , images=_A ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names ) def __lowerCAmelCase ( self : int ) -> str: __magic_name__ : Optional[int] = self.get_image_processor() __magic_name__ : str = self.get_tokenizer() __magic_name__ : List[str] = MgpstrProcessor(tokenizer=_A , image_processor=_A ) __magic_name__ : List[str] = torch.randn(1 , 27 , 38 ) __magic_name__ : Optional[Any] = torch.randn(1 , 27 , 50257 ) __magic_name__ : Dict = torch.randn(1 , 27 , 30522 ) __magic_name__ : Any = processor.batch_decode([char_input, bpe_input, wp_input] ) self.assertListEqual(list(results.keys() ) , ['generated_text', 'scores', 'char_preds', 'bpe_preds', 'wp_preds'] )
275
'''simple docstring''' def lowerCamelCase ( lowerCAmelCase : str ): """simple docstring""" return "".join(chr(ord(lowerCAmelCase ) - 32 ) if 'a' <= char <= 'z' else char for char in word ) if __name__ == "__main__": from doctest import testmod testmod()
275
1
"""simple docstring""" import os import pytest from attr import dataclass a_ = """us-east-1""" # defaults region @dataclass class __snake_case : """simple docstring""" _lowerCamelCase = 42 _lowerCamelCase = """arn:aws:iam::558105141721:role/sagemaker_execution_role""" _lowerCamelCase = { """task_name""": """mnli""", """per_device_train_batch_size""": 16, """per_device_eval_batch_size""": 16, """do_train""": True, """do_eval""": True, """do_predict""": True, """output_dir""": """/opt/ml/model""", """overwrite_output_dir""": True, """max_steps""": 5_00, """save_steps""": 55_00, } _lowerCamelCase = {**hyperparameters, """max_steps""": 10_00} @property def UpperCamelCase__( self ): '''simple docstring''' if self.framework == "pytorch": return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"eval_accuracy.*=\D*(.*?)$"}, {"Name": "eval_loss", "Regex": r"eval_loss.*=\D*(.*?)$"}, ] else: return [ {"Name": "train_runtime", "Regex": r"train_runtime.*=\D*(.*?)$"}, {"Name": "eval_accuracy", "Regex": r"loss.*=\D*(.*?)]?$"}, {"Name": "eval_loss", "Regex": r"sparse_categorical_accuracy.*=\D*(.*?)]?$"}, ] @property def UpperCamelCase__( self ): '''simple docstring''' return F"""{self.framework}-transfromers-test""" @property def UpperCamelCase__( self ): '''simple docstring''' return F"""./tests/sagemaker/scripts/{self.framework}""" @property def UpperCamelCase__( self ): '''simple docstring''' if self.framework == "pytorch": return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-training:1.7.1-transformers4.6.1-gpu-py36-cu110-ubuntu18.04" else: return "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-tensorflow-training:2.4.1-transformers4.6.1-gpu-py37-cu110-ubuntu18.04" @pytest.fixture(scope='''class''' ) def __lowercase ( snake_case_ : int ) ->Tuple: '''simple docstring''' __A : str = SageMakerTestEnvironment(framework=request.cls.framework )
179
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices a_ = logging.get_logger(__name__) a_ = { """shi-labs/nat-mini-in1k-224""": """https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json""", # See all Nat models at https://huggingface.co/models?filter=nat } class __snake_case ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" _lowerCamelCase = """nat""" _lowerCamelCase = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , __lowerCamelCase=4 , __lowerCamelCase=3 , __lowerCamelCase=64 , __lowerCamelCase=[3, 4, 6, 5] , __lowerCamelCase=[2, 4, 8, 16] , __lowerCamelCase=7 , __lowerCamelCase=3.0 , __lowerCamelCase=True , __lowerCamelCase=0.0 , __lowerCamelCase=0.0 , __lowerCamelCase=0.1 , __lowerCamelCase="gelu" , __lowerCamelCase=0.0_2 , __lowerCamelCase=1e-5 , __lowerCamelCase=0.0 , __lowerCamelCase=None , __lowerCamelCase=None , **__lowerCamelCase , ): '''simple docstring''' super().__init__(**__lowerCamelCase ) __A : Union[str, Any] = patch_size __A : Optional[Any] = num_channels __A : Tuple = embed_dim __A : Dict = depths __A : str = len(__lowerCamelCase ) __A : Optional[Any] = num_heads __A : str = kernel_size __A : Any = mlp_ratio __A : Optional[int] = qkv_bias __A : str = hidden_dropout_prob __A : Any = attention_probs_dropout_prob __A : int = drop_path_rate __A : int = hidden_act __A : Any = layer_norm_eps __A : Tuple = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __A : int = int(embed_dim * 2 ** (len(__lowerCamelCase ) - 1) ) __A : Union[str, Any] = layer_scale_init_value __A : List[str] = ['''stem'''] + [F"""stage{idx}""" for idx in range(1 , len(__lowerCamelCase ) + 1 )] __A , __A : Any = get_aligned_output_features_output_indices( out_features=__lowerCamelCase , out_indices=__lowerCamelCase , stage_names=self.stage_names )
179
1
import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging __lowerCAmelCase : Optional[int] ={ 'cola': 2, 'mnli': 3, 'mrpc': 2, 'sst-2': 2, 'sts-b': 1, 'qqp': 2, 'qnli': 2, 'rte': 2, 'wnli': 2, } logging.set_verbosity_info() def UpperCAmelCase__ ( lowerCAmelCase__ :List[str] , lowerCAmelCase__ :Optional[int] , lowerCAmelCase__ :Tuple , lowerCAmelCase__ :Union[str, Any]=None ) -> List[Any]: '''simple docstring''' lowercase = XLNetConfig.from_json_file(_A ) lowercase = finetuning_task.lower() if finetuning_task is not None else '' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) lowercase = finetuning_task lowercase = GLUE_TASKS_NUM_LABELS[finetuning_task] lowercase = XLNetForSequenceClassification(_A ) elif "squad" in finetuning_task: lowercase = finetuning_task lowercase = XLNetForQuestionAnswering(_A ) else: lowercase = XLNetLMHeadModel(_A ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(_A , _A , _A ) # Save pytorch-model lowercase = os.path.join(_A , _A ) lowercase = os.path.join(_A , _A ) print(f'Save PyTorch model to {os.path.abspath(_A )}' ) torch.save(model.state_dict() , _A ) print(f'Save configuration file to {os.path.abspath(_A )}' ) with open(_A , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": __lowerCAmelCase : str =argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--xlnet_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained XLNet model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the folder to store the PyTorch model or dataset/vocab.""", ) parser.add_argument( """--finetuning_task""", default=None, type=str, help="""Name of a task on which the XLNet TensorFlow model was fine-tuned""", ) __lowerCAmelCase : Any =parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
357
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import XLMRobertaTokenizerFast from diffusers import DDIMScheduler, KandinskyInpaintPipeline, KandinskyPriorPipeline, UNetaDConditionModel, VQModel from diffusers.pipelines.kandinsky.text_encoder import MCLIPConfig, MultilingualCLIP from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _A ( lowerCAmelCase , unittest.TestCase ): snake_case__ : str = KandinskyInpaintPipeline snake_case__ : Optional[int] = ['prompt', 'image_embeds', 'negative_image_embeds', 'image', 'mask_image'] snake_case__ : Optional[int] = [ 'prompt', 'negative_prompt', 'image_embeds', 'negative_image_embeds', 'image', 'mask_image', ] snake_case__ : Tuple = [ 'generator', 'height', 'width', 'latents', 'guidance_scale', 'negative_prompt', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] snake_case__ : Dict = False @property def A__ ( self ): """simple docstring""" return 32 @property def A__ ( self ): """simple docstring""" return 32 @property def A__ ( self ): """simple docstring""" return self.time_input_dim @property def A__ ( self ): """simple docstring""" return self.time_input_dim * 4 @property def A__ ( self ): """simple docstring""" return 100 @property def A__ ( self ): """simple docstring""" lowercase = XLMRobertaTokenizerFast.from_pretrained("""YiYiXu/tiny-random-mclip-base""" ) return tokenizer @property def A__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowercase = MCLIPConfig( numDims=self.cross_attention_dim , transformerDimensions=self.text_embedder_hidden_size , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=1005 , ) lowercase = MultilingualCLIP(__lowerCAmelCase ) lowercase = text_encoder.eval() return text_encoder @property def A__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowercase = { """in_channels""": 9, # Out channels is double in channels because predicts mean and variance """out_channels""": 8, """addition_embed_type""": """text_image""", """down_block_types""": ("""ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D"""), """up_block_types""": ("""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""), """mid_block_type""": """UNetMidBlock2DSimpleCrossAttn""", """block_out_channels""": (self.block_out_channels_a, self.block_out_channels_a * 2), """layers_per_block""": 1, """encoder_hid_dim""": self.text_embedder_hidden_size, """encoder_hid_dim_type""": """text_image_proj""", """cross_attention_dim""": self.cross_attention_dim, """attention_head_dim""": 4, """resnet_time_scale_shift""": """scale_shift""", """class_embed_type""": None, } lowercase = UNetaDConditionModel(**__lowerCAmelCase ) return model @property def A__ ( self ): """simple docstring""" return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def A__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowercase = VQModel(**self.dummy_movq_kwargs ) return model def A__ ( self ): """simple docstring""" lowercase = self.dummy_text_encoder lowercase = self.dummy_tokenizer lowercase = self.dummy_unet lowercase = self.dummy_movq lowercase = DDIMScheduler( num_train_timesteps=1000 , beta_schedule="""linear""" , beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , clip_sample=__lowerCAmelCase , set_alpha_to_one=__lowerCAmelCase , steps_offset=1 , prediction_type="""epsilon""" , thresholding=__lowerCAmelCase , ) lowercase = { """text_encoder""": text_encoder, """tokenizer""": tokenizer, """unet""": unet, """scheduler""": scheduler, """movq""": movq, } return components def A__ ( self , __lowerCAmelCase , __lowerCAmelCase=0 ): """simple docstring""" lowercase = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(__lowerCAmelCase ) ).to(__lowerCAmelCase ) lowercase = floats_tensor((1, self.cross_attention_dim) , rng=random.Random(seed + 1 ) ).to(__lowerCAmelCase ) # create init_image lowercase = floats_tensor((1, 3, 64, 64) , rng=random.Random(__lowerCAmelCase ) ).to(__lowerCAmelCase ) lowercase = image.cpu().permute(0 , 2 , 3 , 1 )[0] lowercase = Image.fromarray(np.uinta(__lowerCAmelCase ) ).convert("""RGB""" ).resize((256, 256) ) # create mask lowercase = np.ones((64, 64) , dtype=np.floataa ) lowercase = 0 if str(__lowerCAmelCase ).startswith("""mps""" ): lowercase = torch.manual_seed(__lowerCAmelCase ) else: lowercase = torch.Generator(device=__lowerCAmelCase ).manual_seed(__lowerCAmelCase ) lowercase = { """prompt""": """horse""", """image""": init_image, """mask_image""": mask, """image_embeds""": image_embeds, """negative_image_embeds""": negative_image_embeds, """generator""": generator, """height""": 64, """width""": 64, """num_inference_steps""": 2, """guidance_scale""": 4.0, """output_type""": """np""", } return inputs def A__ ( self ): """simple docstring""" lowercase = """cpu""" lowercase = self.get_dummy_components() lowercase = self.pipeline_class(**__lowerCAmelCase ) lowercase = pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) lowercase = pipe(**self.get_dummy_inputs(__lowerCAmelCase ) ) lowercase = output.images lowercase = pipe( **self.get_dummy_inputs(__lowerCAmelCase ) , return_dict=__lowerCAmelCase , )[0] lowercase = image[0, -3:, -3:, -1] lowercase = image_from_tuple[0, -3:, -3:, -1] print(f'image.shape {image.shape}' ) assert image.shape == (1, 64, 64, 3) lowercase = np.array( [0.8_3_2_6_9_1_9, 0.7_3_7_9_0_4_6_7, 0.2_0_9_1_8_5_8_1, 0.9_3_0_9_6_1_2, 0.5_5_1_1_7_9_1, 0.4_3_7_1_3_3_2_8, 0.5_5_1_3_3_2_1, 0.4_9_9_2_2_9_3_4, 0.5_9_4_9_7_7_8_6] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' def A__ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class _A ( unittest.TestCase ): def A__ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def A__ ( self ): """simple docstring""" lowercase = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/kandinsky_inpaint_cat_with_hat_fp16.npy""" ) lowercase = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/kandinsky/cat.png""" ) lowercase = np.ones((768, 768) , dtype=np.floataa ) lowercase = 0 lowercase = """a hat""" lowercase = KandinskyPriorPipeline.from_pretrained( """kandinsky-community/kandinsky-2-1-prior""" , torch_dtype=torch.floataa ) pipe_prior.to(__lowerCAmelCase ) lowercase = KandinskyInpaintPipeline.from_pretrained( """kandinsky-community/kandinsky-2-1-inpaint""" , torch_dtype=torch.floataa ) lowercase = pipeline.to(__lowerCAmelCase ) pipeline.set_progress_bar_config(disable=__lowerCAmelCase ) lowercase = torch.Generator(device="""cpu""" ).manual_seed(0 ) lowercase , lowercase = pipe_prior( __lowerCAmelCase , generator=__lowerCAmelCase , num_inference_steps=5 , negative_prompt="""""" , ).to_tuple() lowercase = pipeline( __lowerCAmelCase , image=__lowerCAmelCase , mask_image=__lowerCAmelCase , image_embeds=__lowerCAmelCase , negative_image_embeds=__lowerCAmelCase , generator=__lowerCAmelCase , num_inference_steps=100 , height=768 , width=768 , output_type="""np""" , ) lowercase = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowerCAmelCase , __lowerCAmelCase )
32
0
import fire from transformers import AutoConfig, AutoModelForSeqaSeqLM, AutoTokenizer def __lowerCamelCase ( snake_case__ ,snake_case__ ,**snake_case__ ) -> Optional[int]: """simple docstring""" _SCREAMING_SNAKE_CASE = AutoConfig.from_pretrained(__lowerCAmelCase ,**__lowerCAmelCase ) _SCREAMING_SNAKE_CASE = AutoModelForSeqaSeqLM.from_config(__lowerCAmelCase ) model.save_pretrained(__lowerCAmelCase ) AutoTokenizer.from_pretrained(__lowerCAmelCase ).save_pretrained(__lowerCAmelCase ) return model if __name__ == "__main__": fire.Fire(save_randomly_initialized_version)
306
from queue import PriorityQueue from typing import Any import numpy as np def __magic_name__ ( __lowerCAmelCase : dict , __lowerCAmelCase : str , __lowerCAmelCase : set , __lowerCAmelCase : set , __lowerCAmelCase : dict , __lowerCAmelCase : dict , __lowerCAmelCase : PriorityQueue , __lowerCAmelCase : dict , __lowerCAmelCase : float | int , ) -> float | int: for nxt, d in graph[v]: if nxt in visited_forward: continue __lowerCamelCase = cst_fwd.get(__lowerCAmelCase , np.inf ) __lowerCamelCase = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) __lowerCamelCase = new_cost_f __lowerCamelCase = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: __lowerCamelCase = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def __magic_name__ ( __lowerCAmelCase : str , __lowerCAmelCase : str , __lowerCAmelCase : dict , __lowerCAmelCase : dict ) -> int: __lowerCamelCase = -1 __lowerCamelCase = set() __lowerCamelCase = set() __lowerCamelCase = {source: 0} __lowerCamelCase = {destination: 0} __lowerCamelCase = {source: None} __lowerCamelCase = {destination: None} __lowerCamelCase = PriorityQueue() __lowerCamelCase = PriorityQueue() __lowerCamelCase = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): __lowerCamelCase , __lowerCamelCase = queue_forward.get() visited_forward.add(__lowerCAmelCase ) __lowerCamelCase , __lowerCamelCase = queue_backward.get() visited_backward.add(__lowerCAmelCase ) __lowerCamelCase = pass_and_relaxation( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ) __lowerCamelCase = pass_and_relaxation( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: __lowerCamelCase = shortest_distance return shortest_path_distance SCREAMING_SNAKE_CASE__ : List[Any] = { "B": [["C", 1]], "C": [["D", 1]], "D": [["F", 1]], "E": [["B", 1], ["G", 2]], "F": [], "G": [["F", 1]], } SCREAMING_SNAKE_CASE__ : Optional[int] = { "B": [["E", 1]], "C": [["B", 1]], "D": [["C", 1]], "F": [["D", 1], ["G", 1]], "E": [[None, np.inf]], "G": [["E", 2]], } if __name__ == "__main__": import doctest doctest.testmod()
270
0
'''simple docstring''' import datasets from .evaluate import evaluate a_ : List[Any] = """\ @article{hendrycks2021cuad, title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review}, author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball}, journal={arXiv preprint arXiv:2103.06268}, year={2021} } """ a_ : List[Any] = """ This metric wrap the official scoring script for version 1 of the Contract Understanding Atticus Dataset (CUAD). Contract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510 commercial legal contracts that have been manually labeled to identify 41 categories of important clauses that lawyers look for when reviewing contracts in connection with corporate transactions. """ a_ : Any = """ Computes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall). Args: predictions: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair as given in the references (see below) - 'prediction_text': list of possible texts for the answer, as a list of strings depending on a threshold on the confidence probability of each prediction. references: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair (see above), - 'answers': a Dict in the CUAD dataset format { 'text': list of possible texts for the answer, as a list of strings 'answer_start': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: 'exact_match': Exact match (the normalized answer exactly match the gold answer) 'f1': The F-score of predicted tokens versus the gold answer 'aupr': Area Under the Precision-Recall curve 'prec_at_80_recall': Precision at 80% recall 'prec_at_90_recall': Precision at 90% recall Examples: >>> predictions = [{'prediction_text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.'], 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}] >>> references = [{'answers': {'answer_start': [143, 49], 'text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.']}, 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}] >>> cuad_metric = datasets.load_metric(\"cuad\") >>> results = cuad_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 100.0, 'f1': 100.0, 'aupr': 0.0, 'prec_at_80_recall': 1.0, 'prec_at_90_recall': 1.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __UpperCamelCase ( datasets.Metric ): def lowercase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { '''predictions''': { '''id''': datasets.Value('''string''' ), '''prediction_text''': datasets.features.Sequence(datasets.Value('''string''' ) ), }, '''references''': { '''id''': datasets.Value('''string''' ), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string''' ), '''answer_start''': datasets.Value('''int32''' ), } ), }, } ), codebase_urls=['''https://www.atticusprojectai.org/cuad'''], reference_urls=['''https://www.atticusprojectai.org/cuad'''], ) def lowercase__ ( self, lowerCAmelCase, lowerCAmelCase ): """simple docstring""" lowerCamelCase_ ={prediction['''id''']: prediction['''prediction_text'''] for prediction in predictions} lowerCamelCase_ =[ { '''paragraphs''': [ { '''qas''': [ { '''answers''': [{'''text''': answer_text} for answer_text in ref['''answers''']['''text''']], '''id''': ref['''id'''], } for ref in references ] } ] } ] lowerCamelCase_ =evaluate(dataset=lowerCAmelCase, predictions=lowerCAmelCase ) return score
361
'''simple docstring''' def a_ ( __snake_case : int = 1000 ) -> int: """simple docstring""" lowerCamelCase_, lowerCamelCase_ =1, 1 lowerCamelCase_ =2 while True: lowerCamelCase_ =0 lowerCamelCase_ =fa + fa lowerCamelCase_, lowerCamelCase_ =fa, f index += 1 for _ in str(__snake_case ): i += 1 if i == n: break return index if __name__ == "__main__": print(solution(int(str(input()).strip())))
6
0
import argparse from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection from diffusers import UnCLIPImageVariationPipeline, UnCLIPPipeline if __name__ == "__main__": __lowerCAmelCase : List[str] = argparse.ArgumentParser() parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument( '--txt2img_unclip', default='kakaobrain/karlo-v1-alpha', type=str, required=False, help='The pretrained txt2img unclip.', ) __lowerCAmelCase : Optional[int] = parser.parse_args() __lowerCAmelCase : Union[str, Any] = UnCLIPPipeline.from_pretrained(args.txtaimg_unclip) __lowerCAmelCase : List[Any] = CLIPImageProcessor() __lowerCAmelCase : Tuple = CLIPVisionModelWithProjection.from_pretrained('openai/clip-vit-large-patch14') __lowerCAmelCase : Union[str, Any] = UnCLIPImageVariationPipeline( decoder=txtaimg.decoder, text_encoder=txtaimg.text_encoder, tokenizer=txtaimg.tokenizer, text_proj=txtaimg.text_proj, feature_extractor=feature_extractor, image_encoder=image_encoder, super_res_first=txtaimg.super_res_first, super_res_last=txtaimg.super_res_last, decoder_scheduler=txtaimg.decoder_scheduler, super_res_scheduler=txtaimg.super_res_scheduler, ) imgaimg.save_pretrained(args.dump_path)
107
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTImageProcessor, ViTMSNConfig, ViTMSNModel from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD torch.set_grad_enabled(False) def UpperCAmelCase_( a__ , a__=False ): """simple docstring""" SCREAMING_SNAKE_CASE : int = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""module.blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""module.blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (F"""module.blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""module.blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""module.blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""module.blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""module.blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''module.cls_token''', '''vit.embeddings.cls_token'''), ('''module.patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''module.patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''module.pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''module.norm.weight''', '''layernorm.weight'''), ('''module.norm.bias''', '''layernorm.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" SCREAMING_SNAKE_CASE : Any = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def UpperCAmelCase_( a__ , a__ , a__=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: SCREAMING_SNAKE_CASE : Any = '''''' else: SCREAMING_SNAKE_CASE : Optional[int] = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop(F"""module.blocks.{i}.attn.qkv.weight""" ) SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(F"""module.blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : List[str] = in_proj_weight[ : config.hidden_size, : ] SCREAMING_SNAKE_CASE : Any = in_proj_bias[: config.hidden_size] SCREAMING_SNAKE_CASE : int = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] SCREAMING_SNAKE_CASE : List[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] SCREAMING_SNAKE_CASE : List[str] = in_proj_weight[ -config.hidden_size :, : ] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_bias[-config.hidden_size :] def UpperCAmelCase_( a__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(a__ , a__ ) def UpperCAmelCase_( a__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = [ '''module.fc.fc1.weight''', '''module.fc.fc1.bias''', '''module.fc.bn1.weight''', '''module.fc.bn1.bias''', '''module.fc.bn1.running_mean''', '''module.fc.bn1.running_var''', '''module.fc.bn1.num_batches_tracked''', '''module.fc.fc2.weight''', '''module.fc.fc2.bias''', '''module.fc.bn2.weight''', '''module.fc.bn2.bias''', '''module.fc.bn2.running_mean''', '''module.fc.bn2.running_var''', '''module.fc.bn2.num_batches_tracked''', '''module.fc.fc3.weight''', '''module.fc.fc3.bias''', ] for k in ignore_keys: state_dict.pop(a__ , a__ ) def UpperCAmelCase_( a__ , a__ , a__ ): """simple docstring""" SCREAMING_SNAKE_CASE : int = dct.pop(a__ ) SCREAMING_SNAKE_CASE : Optional[Any] = val def UpperCAmelCase_( a__ , a__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = ViTMSNConfig() SCREAMING_SNAKE_CASE : Optional[int] = 1_000 SCREAMING_SNAKE_CASE : str = '''datasets/huggingface/label-files''' SCREAMING_SNAKE_CASE : List[str] = '''imagenet-1k-id2label.json''' SCREAMING_SNAKE_CASE : Union[str, Any] = json.load(open(hf_hub_download(a__ , a__ ) , '''r''' ) ) SCREAMING_SNAKE_CASE : List[Any] = {int(a__ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : str = idalabel SCREAMING_SNAKE_CASE : int = {v: k for k, v in idalabel.items()} if "s16" in checkpoint_url: SCREAMING_SNAKE_CASE : Tuple = 384 SCREAMING_SNAKE_CASE : Any = 1_536 SCREAMING_SNAKE_CASE : List[str] = 6 elif "l16" in checkpoint_url: SCREAMING_SNAKE_CASE : Optional[int] = 1_024 SCREAMING_SNAKE_CASE : Optional[int] = 4_096 SCREAMING_SNAKE_CASE : Tuple = 24 SCREAMING_SNAKE_CASE : Union[str, Any] = 16 SCREAMING_SNAKE_CASE : Dict = 0.1 elif "b4" in checkpoint_url: SCREAMING_SNAKE_CASE : str = 4 elif "l7" in checkpoint_url: SCREAMING_SNAKE_CASE : Union[str, Any] = 7 SCREAMING_SNAKE_CASE : Union[str, Any] = 1_024 SCREAMING_SNAKE_CASE : List[Any] = 4_096 SCREAMING_SNAKE_CASE : List[Any] = 24 SCREAMING_SNAKE_CASE : Tuple = 16 SCREAMING_SNAKE_CASE : Union[str, Any] = 0.1 SCREAMING_SNAKE_CASE : Union[str, Any] = ViTMSNModel(a__ ) SCREAMING_SNAKE_CASE : Optional[int] = torch.hub.load_state_dict_from_url(a__ , map_location='''cpu''' )['''target_encoder'''] SCREAMING_SNAKE_CASE : Any = ViTImageProcessor(size=config.image_size ) remove_projection_head(a__ ) SCREAMING_SNAKE_CASE : Any = create_rename_keys(a__ , base_model=a__ ) for src, dest in rename_keys: rename_key(a__ , a__ , a__ ) read_in_q_k_v(a__ , a__ , base_model=a__ ) model.load_state_dict(a__ ) model.eval() SCREAMING_SNAKE_CASE : Any = '''http://images.cocodataset.org/val2017/000000039769.jpg''' SCREAMING_SNAKE_CASE : Dict = Image.open(requests.get(a__ , stream=a__ ).raw ) SCREAMING_SNAKE_CASE : Optional[int] = ViTImageProcessor( size=config.image_size , image_mean=a__ , image_std=a__ ) SCREAMING_SNAKE_CASE : int = image_processor(images=a__ , return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) SCREAMING_SNAKE_CASE : Tuple = model(**a__ ) SCREAMING_SNAKE_CASE : str = outputs.last_hidden_state # The following Colab Notebook was used to generate these outputs: # https://colab.research.google.com/gist/sayakpaul/3672419a04f5997827503fd84079bdd1/scratchpad.ipynb if "s16" in checkpoint_url: SCREAMING_SNAKE_CASE : Tuple = torch.tensor([[-1.0_915, -1.4_876, -1.1_809]] ) elif "b16" in checkpoint_url: SCREAMING_SNAKE_CASE : Tuple = torch.tensor([[14.2_889, -18.9_045, 11.7_281]] ) elif "l16" in checkpoint_url: SCREAMING_SNAKE_CASE : Union[str, Any] = torch.tensor([[41.5_028, -22.8_681, 45.6_475]] ) elif "b4" in checkpoint_url: SCREAMING_SNAKE_CASE : str = torch.tensor([[-4.3_868, 5.2_932, -0.4_137]] ) else: SCREAMING_SNAKE_CASE : List[str] = torch.tensor([[-0.1_792, -0.6_465, 2.4_263]] ) # verify logits assert torch.allclose(last_hidden_state[:, 0, :3] , a__ , atol=1e-4 ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(a__ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a__ ) if __name__ == "__main__": a__ : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://dl.fbaipublicfiles.com/msn/vits16_800ep.pth.tar''', type=str, help='''URL of the checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) a__ : Any = parser.parse_args() convert_vit_msn_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
313
0
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __magic_name__ ( unittest.TestCase ): @property def __lowercase ( self : List[str] ): torch.manual_seed(0 ) _a : Union[str, Any] = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=('DownBlock2D', 'AttnDownBlock2D') ,up_block_types=('AttnUpBlock2D', 'UpBlock2D') ,) return model @property def __lowercase ( self : List[str] ): torch.manual_seed(0 ) _a : Tuple = VQModel( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] ,up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] ,latent_channels=3 ,) return model @property def __lowercase ( self : List[str] ): torch.manual_seed(0 ) _a : Optional[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1E-05 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=1000 ,) return CLIPTextModel(_UpperCAmelCase ) def __lowercase ( self : Union[str, Any] ): _a : Optional[Any] = self.dummy_uncond_unet _a : int = DDIMScheduler() _a : Optional[int] = self.dummy_vq_model _a : List[Any] = LDMPipeline(unet=_UpperCAmelCase ,vqvae=_UpperCAmelCase ,scheduler=_UpperCAmelCase ) ldm.to(_UpperCAmelCase ) ldm.set_progress_bar_config(disable=_UpperCAmelCase ) _a : Dict = torch.manual_seed(0 ) _a : Optional[int] = ldm(generator=_UpperCAmelCase ,num_inference_steps=2 ,output_type='numpy' ).images _a : Union[str, Any] = torch.manual_seed(0 ) _a : Union[str, Any] = ldm(generator=_UpperCAmelCase ,num_inference_steps=2 ,output_type='numpy' ,return_dict=_UpperCAmelCase )[0] _a : Optional[int] = image[0, -3:, -3:, -1] _a : int = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _a : Optional[Any] = np.array([0.85_12, 0.8_18, 0.64_11, 0.68_08, 0.44_65, 0.56_18, 0.46, 0.62_31, 0.51_72] ) _a : Tuple = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class __magic_name__ ( unittest.TestCase ): def __lowercase ( self : List[Any] ): _a : Any = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(_UpperCAmelCase ) ldm.set_progress_bar_config(disable=_UpperCAmelCase ) _a : Any = torch.manual_seed(0 ) _a : Any = ldm(generator=_UpperCAmelCase ,num_inference_steps=5 ,output_type='numpy' ).images _a : str = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) _a : Any = np.array([0.43_99, 0.4_49_75, 0.4_68_25, 0.4_74, 0.43_59, 0.45_81, 0.4_50_95, 0.43_41, 0.44_47] ) _a : int = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
107
'''simple docstring''' import heapq def __lowerCamelCase ( lowerCAmelCase_ ) -> set[int]: _a : list[list] = [] # for each node and his adjacency list add them and the rank of the node to queue # using heapq module the queue will be filled like a Priority Queue # heapq works with a min priority queue, so I used -1*len(v) to build it for key, value in graph.items(): # O(log(n)) heapq.heappush(lowerCAmelCase_ , [-1 * len(lowerCAmelCase_ ), (key, value)] ) # chosen_vertices = set of chosen vertices _a : str = set() # while queue isn't empty and there are still edges # (queue[0][0] is the rank of the node with max rank) while queue and queue[0][0] != 0: # extract vertex with max rank from queue and add it to chosen_vertices _a : Optional[int] = heapq.heappop(lowerCAmelCase_ )[1][0] chosen_vertices.add(lowerCAmelCase_ ) # Remove all arcs adjacent to argmax for elem in queue: # if v haven't adjacent node, skip if elem[0] == 0: continue # if argmax is reachable from elem # remove argmax from elem's adjacent list and update his rank if argmax in elem[1][1]: _a : Dict = elem[1][1].index(lowerCAmelCase_ ) del elem[1][1][index] elem[0] += 1 # re-order the queue heapq.heapify(lowerCAmelCase_ ) return chosen_vertices if __name__ == "__main__": import doctest doctest.testmod() __lowerCAmelCase = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} print(f"""Minimum vertex cover:\n{greedy_min_vertex_cover(graph)}""")
107
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: lowercase__ : str = None lowercase__ : Tuple = logging.get_logger(__name__) lowercase__ : List[str] = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} lowercase__ : Optional[int] = { 'vocab_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/tokenizer.json', }, } lowercase__ : str = { 'camembert-base': 5_12, } lowercase__ : Union[str, Any] = '▁' class __lowerCAmelCase ( _lowercase ): """simple docstring""" _snake_case : Any = VOCAB_FILES_NAMES _snake_case : List[Any] = PRETRAINED_VOCAB_FILES_MAP _snake_case : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : Optional[Any] = ['input_ids', 'attention_mask'] _snake_case : Dict = CamembertTokenizer def __init__( self : Dict , lowerCAmelCase__ : Union[str, Any]=None , lowerCAmelCase__ : Tuple=None , lowerCAmelCase__ : Any="<s>" , lowerCAmelCase__ : Dict="</s>" , lowerCAmelCase__ : Any="</s>" , lowerCAmelCase__ : str="<s>" , lowerCAmelCase__ : Any="<unk>" , lowerCAmelCase__ : Union[str, Any]="<pad>" , lowerCAmelCase__ : List[str]="<mask>" , lowerCAmelCase__ : List[Any]=["<s>NOTUSED", "</s>NOTUSED"] , **lowerCAmelCase__ : Dict , ) -> str: '''simple docstring''' _UpperCamelCase = AddedToken(__a , lstrip=__a , rstrip=__a ) if isinstance(__a , __a ) else mask_token super().__init__( __a , tokenizer_file=__a , bos_token=__a , eos_token=__a , sep_token=__a , cls_token=__a , unk_token=__a , pad_token=__a , mask_token=__a , additional_special_tokens=__a , **__a , ) _UpperCamelCase = vocab_file _UpperCamelCase = False if not self.vocab_file else True def snake_case__ ( self : List[Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : Union[str, Any] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] _UpperCamelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def snake_case__ ( self : Optional[int] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : List[Any] = None ) -> List[int]: '''simple docstring''' _UpperCamelCase = [self.sep_token_id] _UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self : str , lowerCAmelCase__ : str , lowerCAmelCase__ : Dict = None ) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _UpperCamelCase = os.path.join( __a , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__a ): copyfile(self.vocab_file , __a ) return (out_vocab_file,)
324
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline lowerCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name class _lowerCamelCase ( _lowercase ): def __init__(self , __a , __a ) -> str: super().__init__() self.register_modules(unet=__a , scheduler=__a ) @torch.no_grad() def __call__(self , __a = 1 , __a = 1_00 , __a = None , __a = None , __a = True , ) -> Union[AudioPipelineOutput, Tuple]: if audio_length_in_s is None: UpperCamelCase = self.unet.config.sample_size / self.unet.config.sample_rate UpperCamelCase = audio_length_in_s * self.unet.config.sample_rate UpperCamelCase = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( F"{audio_length_in_s} is too small. Make sure it's bigger or equal to" F" {3 * down_scale_factor / self.unet.config.sample_rate}." ) UpperCamelCase = int(__a ) if sample_size % down_scale_factor != 0: UpperCamelCase = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( F"{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled" F" by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising" " process." ) UpperCamelCase = int(__a ) UpperCamelCase = next(iter(self.unet.parameters() ) ).dtype UpperCamelCase = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(__a , __a ) and len(__a ) != batch_size: raise ValueError( F"You have passed a list of generators of length {len(__a )}, but requested an effective batch" F" size of {batch_size}. Make sure the batch size matches the length of the generators." ) UpperCamelCase = randn_tensor(__a , generator=__a , device=self.device , dtype=__a ) # set step values self.scheduler.set_timesteps(__a , device=audio.device ) UpperCamelCase = self.scheduler.timesteps.to(__a ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output UpperCamelCase = self.unet(__a , __a ).sample # 2. compute previous image: x_t -> t_t-1 UpperCamelCase = self.scheduler.step(__a , __a , __a ).prev_sample UpperCamelCase = audio.clamp(-1 , 1 ).float().cpu().numpy() UpperCamelCase = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=__a )
153
0
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging __A = logging.get_logger(__name__) __A = R"\n Args:\n input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax\n or scores for each vocabulary token after SoftMax.\n kwargs (`Dict[str, Any]`, *optional*):\n Additional stopping criteria specific kwargs.\n\n Return:\n `bool`. `False` indicates we should continue, `True` indicates we should stop.\n\n" class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" @add_start_docstrings(_UpperCAmelCase ) def __call__( self , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ): raise NotImplementedError('''StoppingCriteria needs to be subclassed''' ) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase = None ): lowercase__: Union[str, Any] = max_length lowercase__: str = max_position_embeddings @add_start_docstrings(_UpperCAmelCase ) def __call__( self , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ): lowercase__: Any = input_ids.shape[-1] lowercase__: str = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( '''This is a friendly reminder - the current text generation call will exceed the model\'s predefined ''' F"""maximum length ({self.max_position_embeddings}). Depending on the model, you may observe """ '''exceptions, performance degradation, or nothing at all.''' ) return is_done class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase ): warnings.warn( '''The class `MaxNewTokensCriteria` is deprecated. ''' F"""Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` """ '''with `max_length = start_length + max_new_tokens` instead.''' , _UpperCAmelCase , ) lowercase__: Optional[Any] = start_length lowercase__: Union[str, Any] = max_new_tokens lowercase__: int = start_length + max_new_tokens @add_start_docstrings(_UpperCAmelCase ) def __call__( self , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ): return input_ids.shape[-1] >= self.max_length class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase = None ): lowercase__: int = max_time lowercase__: int = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(_UpperCAmelCase ) def __call__( self , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ): return time.time() - self.initial_timestamp > self.max_time class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" @add_start_docstrings(_UpperCAmelCase ) def __call__( self , _UpperCAmelCase , _UpperCAmelCase , **_UpperCAmelCase ): return any(criteria(_UpperCAmelCase , _UpperCAmelCase ) for criteria in self ) @property def _snake_case ( self ): for stopping_criterium in self: if isinstance(_UpperCAmelCase , _UpperCAmelCase ): return stopping_criterium.max_length elif isinstance(_UpperCAmelCase , _UpperCAmelCase ): return stopping_criterium.max_length return None def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase ) -> StoppingCriteriaList: lowercase__: List[str] = stopping_criteria.max_length lowercase__: Optional[int] = deepcopy(__UpperCAmelCase ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn('''You set different `max_length` for stopping criteria and `max_length` parameter''' , __UpperCAmelCase ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=__UpperCAmelCase ) ) return new_stopping_criteria
368
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging __A = logging.get_logger(__name__) __A = { "bigscience/bloom": "https://huggingface.co/bigscience/bloom/resolve/main/config.json", "bigscience/bloom-560m": "https://huggingface.co/bigscience/bloom-560m/blob/main/config.json", "bigscience/bloom-1b1": "https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json", "bigscience/bloom-1b7": "https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json", "bigscience/bloom-3b": "https://huggingface.co/bigscience/bloom-3b/blob/main/config.json", "bigscience/bloom-7b1": "https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json", } class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :str = "bloom" _UpperCAmelCase :List[str] = ["past_key_values"] _UpperCAmelCase :Optional[Any] = { "num_hidden_layers": "n_layer", "num_attention_heads": "n_head", } def __init__( self , _UpperCAmelCase=250880 , _UpperCAmelCase=64 , _UpperCAmelCase=2 , _UpperCAmelCase=8 , _UpperCAmelCase=1e-5 , _UpperCAmelCase=0.02 , _UpperCAmelCase=True , _UpperCAmelCase=1 , _UpperCAmelCase=2 , _UpperCAmelCase=False , _UpperCAmelCase=0.0 , _UpperCAmelCase=0.0 , _UpperCAmelCase=1 , _UpperCAmelCase=False , **_UpperCAmelCase , ): lowercase__: Any = vocab_size # Backward compatibility with n_embed kwarg lowercase__: Optional[Any] = kwargs.pop('''n_embed''' , _UpperCAmelCase ) lowercase__: int = hidden_size if n_embed is None else n_embed lowercase__: int = n_layer lowercase__: int = n_head lowercase__: Optional[Any] = layer_norm_epsilon lowercase__: int = initializer_range lowercase__: List[Any] = use_cache lowercase__: str = pretraining_tp lowercase__: Tuple = apply_residual_connection_post_layernorm lowercase__: int = hidden_dropout lowercase__: Optional[Any] = attention_dropout lowercase__: int = bos_token_id lowercase__: Union[str, Any] = eos_token_id lowercase__: Any = slow_but_exact super().__init__(bos_token_id=_UpperCAmelCase , eos_token_id=_UpperCAmelCase , **_UpperCAmelCase ) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :int = version.parse("1.12" ) def __init__( self , _UpperCAmelCase , _UpperCAmelCase = "default" , _UpperCAmelCase = None , _UpperCAmelCase = False , ): super().__init__(_UpperCAmelCase , task=_UpperCAmelCase , patching_specs=_UpperCAmelCase , use_past=_UpperCAmelCase ) if not getattr(self._config , '''pad_token_id''' , _UpperCAmelCase ): # TODO: how to do that better? lowercase__: Any = 0 @property def _snake_case ( self ): lowercase__: str = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(_UpperCAmelCase , direction='''inputs''' , inverted_values_shape=_UpperCAmelCase ) lowercase__: List[str] = {0: '''batch''', 1: '''past_sequence + sequence'''} else: lowercase__: str = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def _snake_case ( self ): return self._config.n_layer @property def _snake_case ( self ): return self._config.n_head @property def _snake_case ( self ): return 1e-3 def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase = -1 , _UpperCAmelCase = -1 , _UpperCAmelCase = False , _UpperCAmelCase = None , ): lowercase__: str = super(_UpperCAmelCase , self ).generate_dummy_inputs( _UpperCAmelCase , batch_size=_UpperCAmelCase , seq_length=_UpperCAmelCase , is_pair=_UpperCAmelCase , framework=_UpperCAmelCase ) # We need to order the input in the way they appears in the forward() lowercase__: List[Any] = OrderedDict({'''input_ids''': common_inputs['''input_ids''']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch lowercase__, lowercase__: Optional[Any] = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values lowercase__: Tuple = seqlen + 2 lowercase__: str = self._config.hidden_size // self.num_attention_heads lowercase__: Optional[int] = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) lowercase__: Union[str, Any] = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) lowercase__: str = [ (torch.zeros(_UpperCAmelCase ), torch.zeros(_UpperCAmelCase )) for _ in range(self.num_layers ) ] lowercase__: Tuple = common_inputs['''attention_mask'''] if self.use_past: lowercase__: int = ordered_inputs['''attention_mask'''].dtype lowercase__: List[str] = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(_UpperCAmelCase , _UpperCAmelCase , dtype=_UpperCAmelCase )] , dim=1 ) return ordered_inputs @property def _snake_case ( self ): return 13
2
0
from __future__ import annotations import math def UpperCAmelCase_ ( __lowerCAmelCase ) -> bool: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__lowerCAmelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def UpperCAmelCase_ ( __lowerCAmelCase ) -> list[int]: __lowercase : Dict = str(__lowerCAmelCase ) __lowercase : Optional[int] = [n] for i in range(1 , len(__lowerCAmelCase ) ): list_nums.append(int(str_num[i:] ) ) list_nums.append(int(str_num[:-i] ) ) return list_nums def UpperCAmelCase_ ( __lowerCAmelCase ) -> bool: if len(str(__lowerCAmelCase ) ) > 3: if not is_prime(int(str(__lowerCAmelCase )[-3:] ) ) or not is_prime(int(str(__lowerCAmelCase )[:3] ) ): return False return True def UpperCAmelCase_ ( __lowerCAmelCase = 11 ) -> list[int]: __lowercase : list[int] = [] __lowercase : List[str] = 13 while len(__lowerCAmelCase ) != count: if validate(__lowerCAmelCase ): __lowercase : Dict = list_truncated_nums(__lowerCAmelCase ) if all(is_prime(__lowerCAmelCase ) for i in list_nums ): list_truncated_primes.append(__lowerCAmelCase ) num += 2 return list_truncated_primes def UpperCAmelCase_ ( ) -> int: return sum(compute_truncated_primes(11 ) ) if __name__ == "__main__": print(F'{sum(compute_truncated_primes(11)) = }')
156
from collections import deque from .hash_table import HashTable class __lowerCAmelCase ( lowerCAmelCase_ ): """simple docstring""" def __init__( self : Union[str, Any] , *_snake_case : Union[str, Any] , **_snake_case : Union[str, Any] ): super().__init__(*_snake_case , **_snake_case ) def snake_case_ ( self : List[Any] , _snake_case : List[Any] , _snake_case : Dict ): __lowercase : Any = deque([] ) if self.values[key] is None else self.values[key] self.values[key].appendleft(_snake_case ) __lowercase : List[Any] = self.values[key] def snake_case_ ( self : Any ): return ( sum(self.charge_factor - len(_snake_case ) for slot in self.values ) / self.size_table * self.charge_factor ) def snake_case_ ( self : int , _snake_case : str , _snake_case : Optional[int]=None ): if not ( len(self.values[key] ) == self.charge_factor and self.values.count(_snake_case ) == 0 ): return key return super()._collision_resolution(_snake_case , _snake_case )
156
1
from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract snake_case_ = logging.get_logger(__name__) def snake_case__ ( SCREAMING_SNAKE_CASE_ : Any , SCREAMING_SNAKE_CASE_ : Optional[int] , SCREAMING_SNAKE_CASE_ : Union[str, Any] ): '''simple docstring''' return [ int(1_000 * (box[0] / width) ), int(1_000 * (box[1] / height) ), int(1_000 * (box[2] / width) ), int(1_000 * (box[3] / height) ), ] def snake_case__ ( SCREAMING_SNAKE_CASE_ : np.ndarray , SCREAMING_SNAKE_CASE_ : Optional[str] , SCREAMING_SNAKE_CASE_ : Optional[str] = None ): '''simple docstring''' lowercase__ : int = tesseract_config if tesseract_config is not None else '' # apply OCR lowercase__ : List[str] = to_pil_image(_UpperCAmelCase ) lowercase__ : str = pil_image.size lowercase__ : Optional[Any] = pytesseract.image_to_data(_UpperCAmelCase , lang=_UpperCAmelCase , output_type='dict' , config=_UpperCAmelCase ) lowercase__ : Tuple = data['text'], data['left'], data['top'], data['width'], data['height'] # filter empty words and corresponding coordinates lowercase__ : Dict = [idx for idx, word in enumerate(_UpperCAmelCase ) if not word.strip()] lowercase__ : Tuple = [word for idx, word in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase__ : Any = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase__ : int = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase__ : Union[str, Any] = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] lowercase__ : List[Any] = [coord for idx, coord in enumerate(_UpperCAmelCase ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format lowercase__ : Tuple = [] for x, y, w, h in zip(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): lowercase__ : Any = [x, y, x + w, y + h] actual_boxes.append(_UpperCAmelCase ) # finally, normalize the bounding boxes lowercase__ : Optional[Any] = [] for box in actual_boxes: normalized_boxes.append(normalize_box(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) ) assert len(_UpperCAmelCase ) == len(_UpperCAmelCase ), "Not as many words as there are bounding boxes" return words, normalized_boxes class SCREAMING_SNAKE_CASE__ (SCREAMING_SNAKE_CASE_ ): __lowerCamelCase : Tuple = ["""pixel_values"""] def __init__( self , a = True , a = None , a = PILImageResampling.BILINEAR , a = True , a = None , a = "" , **a , ): super().__init__(**__a) lowercase__ : List[Any] = size if size is not None else {'height': 224, 'width': 224} lowercase__ : List[Any] = get_size_dict(__a) lowercase__ : List[Any] = do_resize lowercase__ : Dict = size lowercase__ : Dict = resample lowercase__ : Dict = apply_ocr lowercase__ : List[Any] = ocr_lang lowercase__ : List[str] = tesseract_config def snake_case_ ( self , a , a , a = PILImageResampling.BILINEAR , a = None , **a , ): lowercase__ : Union[str, Any] = get_size_dict(__a) if "height" not in size or "width" not in size: raise ValueError(f"""The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}""") lowercase__ : Dict = (size['height'], size['width']) return resize(__a , size=__a , resample=__a , data_format=__a , **__a) def snake_case_ ( self , a , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = ChannelDimension.FIRST , **a , ): lowercase__ : int = do_resize if do_resize is not None else self.do_resize lowercase__ : Optional[int] = size if size is not None else self.size lowercase__ : Tuple = get_size_dict(__a) lowercase__ : List[Any] = resample if resample is not None else self.resample lowercase__ : Union[str, Any] = apply_ocr if apply_ocr is not None else self.apply_ocr lowercase__ : Any = ocr_lang if ocr_lang is not None else self.ocr_lang lowercase__ : Tuple = tesseract_config if tesseract_config is not None else self.tesseract_config lowercase__ : Tuple = make_list_of_images(__a) if not valid_images(__a): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.') if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.') # All transformations expect numpy arrays. lowercase__ : List[Any] = [to_numpy_array(__a) for image in images] if apply_ocr: requires_backends(self , 'pytesseract') lowercase__ : List[Any] = [] lowercase__ : Union[str, Any] = [] for image in images: lowercase__ : Optional[Any] = apply_tesseract(__a , __a , __a) words_batch.append(__a) boxes_batch.append(__a) if do_resize: lowercase__ : Any = [self.resize(image=__a , size=__a , resample=__a) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) lowercase__ : List[Any] = [flip_channel_order(__a) for image in images] lowercase__ : Dict = [to_channel_dimension_format(__a , __a) for image in images] lowercase__ : Optional[Any] = BatchFeature(data={'pixel_values': images} , tensor_type=__a) if apply_ocr: lowercase__ : int = words_batch lowercase__ : List[str] = boxes_batch return data
350
from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES snake_case_ = logging.get_logger(__name__) snake_case_ = OrderedDict( [ # Base model mapping ('''albert''', '''FlaxAlbertModel'''), ('''bart''', '''FlaxBartModel'''), ('''beit''', '''FlaxBeitModel'''), ('''bert''', '''FlaxBertModel'''), ('''big_bird''', '''FlaxBigBirdModel'''), ('''blenderbot''', '''FlaxBlenderbotModel'''), ('''blenderbot-small''', '''FlaxBlenderbotSmallModel'''), ('''clip''', '''FlaxCLIPModel'''), ('''distilbert''', '''FlaxDistilBertModel'''), ('''electra''', '''FlaxElectraModel'''), ('''gpt-sw3''', '''FlaxGPT2Model'''), ('''gpt2''', '''FlaxGPT2Model'''), ('''gpt_neo''', '''FlaxGPTNeoModel'''), ('''gptj''', '''FlaxGPTJModel'''), ('''longt5''', '''FlaxLongT5Model'''), ('''marian''', '''FlaxMarianModel'''), ('''mbart''', '''FlaxMBartModel'''), ('''mt5''', '''FlaxMT5Model'''), ('''opt''', '''FlaxOPTModel'''), ('''pegasus''', '''FlaxPegasusModel'''), ('''regnet''', '''FlaxRegNetModel'''), ('''resnet''', '''FlaxResNetModel'''), ('''roberta''', '''FlaxRobertaModel'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormModel'''), ('''roformer''', '''FlaxRoFormerModel'''), ('''t5''', '''FlaxT5Model'''), ('''vision-text-dual-encoder''', '''FlaxVisionTextDualEncoderModel'''), ('''vit''', '''FlaxViTModel'''), ('''wav2vec2''', '''FlaxWav2Vec2Model'''), ('''whisper''', '''FlaxWhisperModel'''), ('''xglm''', '''FlaxXGLMModel'''), ('''xlm-roberta''', '''FlaxXLMRobertaModel'''), ] ) snake_case_ = OrderedDict( [ # Model for pre-training mapping ('''albert''', '''FlaxAlbertForPreTraining'''), ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''bert''', '''FlaxBertForPreTraining'''), ('''big_bird''', '''FlaxBigBirdForPreTraining'''), ('''electra''', '''FlaxElectraForPreTraining'''), ('''longt5''', '''FlaxLongT5ForConditionalGeneration'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''mt5''', '''FlaxMT5ForConditionalGeneration'''), ('''roberta''', '''FlaxRobertaForMaskedLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMaskedLM'''), ('''roformer''', '''FlaxRoFormerForMaskedLM'''), ('''t5''', '''FlaxT5ForConditionalGeneration'''), ('''wav2vec2''', '''FlaxWav2Vec2ForPreTraining'''), ('''whisper''', '''FlaxWhisperForConditionalGeneration'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMaskedLM'''), ] ) snake_case_ = OrderedDict( [ # Model for Masked LM mapping ('''albert''', '''FlaxAlbertForMaskedLM'''), ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''bert''', '''FlaxBertForMaskedLM'''), ('''big_bird''', '''FlaxBigBirdForMaskedLM'''), ('''distilbert''', '''FlaxDistilBertForMaskedLM'''), ('''electra''', '''FlaxElectraForMaskedLM'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''roberta''', '''FlaxRobertaForMaskedLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMaskedLM'''), ('''roformer''', '''FlaxRoFormerForMaskedLM'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMaskedLM'''), ] ) snake_case_ = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''blenderbot''', '''FlaxBlenderbotForConditionalGeneration'''), ('''blenderbot-small''', '''FlaxBlenderbotSmallForConditionalGeneration'''), ('''encoder-decoder''', '''FlaxEncoderDecoderModel'''), ('''longt5''', '''FlaxLongT5ForConditionalGeneration'''), ('''marian''', '''FlaxMarianMTModel'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''mt5''', '''FlaxMT5ForConditionalGeneration'''), ('''pegasus''', '''FlaxPegasusForConditionalGeneration'''), ('''t5''', '''FlaxT5ForConditionalGeneration'''), ] ) snake_case_ = OrderedDict( [ # Model for Image-classsification ('''beit''', '''FlaxBeitForImageClassification'''), ('''regnet''', '''FlaxRegNetForImageClassification'''), ('''resnet''', '''FlaxResNetForImageClassification'''), ('''vit''', '''FlaxViTForImageClassification'''), ] ) snake_case_ = OrderedDict( [ ('''vision-encoder-decoder''', '''FlaxVisionEncoderDecoderModel'''), ] ) snake_case_ = OrderedDict( [ # Model for Causal LM mapping ('''bart''', '''FlaxBartForCausalLM'''), ('''bert''', '''FlaxBertForCausalLM'''), ('''big_bird''', '''FlaxBigBirdForCausalLM'''), ('''electra''', '''FlaxElectraForCausalLM'''), ('''gpt-sw3''', '''FlaxGPT2LMHeadModel'''), ('''gpt2''', '''FlaxGPT2LMHeadModel'''), ('''gpt_neo''', '''FlaxGPTNeoForCausalLM'''), ('''gptj''', '''FlaxGPTJForCausalLM'''), ('''opt''', '''FlaxOPTForCausalLM'''), ('''roberta''', '''FlaxRobertaForCausalLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForCausalLM'''), ('''xglm''', '''FlaxXGLMForCausalLM'''), ('''xlm-roberta''', '''FlaxXLMRobertaForCausalLM'''), ] ) snake_case_ = OrderedDict( [ # Model for Sequence Classification mapping ('''albert''', '''FlaxAlbertForSequenceClassification'''), ('''bart''', '''FlaxBartForSequenceClassification'''), ('''bert''', '''FlaxBertForSequenceClassification'''), ('''big_bird''', '''FlaxBigBirdForSequenceClassification'''), ('''distilbert''', '''FlaxDistilBertForSequenceClassification'''), ('''electra''', '''FlaxElectraForSequenceClassification'''), ('''mbart''', '''FlaxMBartForSequenceClassification'''), ('''roberta''', '''FlaxRobertaForSequenceClassification'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForSequenceClassification'''), ('''roformer''', '''FlaxRoFormerForSequenceClassification'''), ('''xlm-roberta''', '''FlaxXLMRobertaForSequenceClassification'''), ] ) snake_case_ = OrderedDict( [ # Model for Question Answering mapping ('''albert''', '''FlaxAlbertForQuestionAnswering'''), ('''bart''', '''FlaxBartForQuestionAnswering'''), ('''bert''', '''FlaxBertForQuestionAnswering'''), ('''big_bird''', '''FlaxBigBirdForQuestionAnswering'''), ('''distilbert''', '''FlaxDistilBertForQuestionAnswering'''), ('''electra''', '''FlaxElectraForQuestionAnswering'''), ('''mbart''', '''FlaxMBartForQuestionAnswering'''), ('''roberta''', '''FlaxRobertaForQuestionAnswering'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForQuestionAnswering'''), ('''roformer''', '''FlaxRoFormerForQuestionAnswering'''), ('''xlm-roberta''', '''FlaxXLMRobertaForQuestionAnswering'''), ] ) snake_case_ = OrderedDict( [ # Model for Token Classification mapping ('''albert''', '''FlaxAlbertForTokenClassification'''), ('''bert''', '''FlaxBertForTokenClassification'''), ('''big_bird''', '''FlaxBigBirdForTokenClassification'''), ('''distilbert''', '''FlaxDistilBertForTokenClassification'''), ('''electra''', '''FlaxElectraForTokenClassification'''), ('''roberta''', '''FlaxRobertaForTokenClassification'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForTokenClassification'''), ('''roformer''', '''FlaxRoFormerForTokenClassification'''), ('''xlm-roberta''', '''FlaxXLMRobertaForTokenClassification'''), ] ) snake_case_ = OrderedDict( [ # Model for Multiple Choice mapping ('''albert''', '''FlaxAlbertForMultipleChoice'''), ('''bert''', '''FlaxBertForMultipleChoice'''), ('''big_bird''', '''FlaxBigBirdForMultipleChoice'''), ('''distilbert''', '''FlaxDistilBertForMultipleChoice'''), ('''electra''', '''FlaxElectraForMultipleChoice'''), ('''roberta''', '''FlaxRobertaForMultipleChoice'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMultipleChoice'''), ('''roformer''', '''FlaxRoFormerForMultipleChoice'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMultipleChoice'''), ] ) snake_case_ = OrderedDict( [ ('''bert''', '''FlaxBertForNextSentencePrediction'''), ] ) snake_case_ = OrderedDict( [ ('''speech-encoder-decoder''', '''FlaxSpeechEncoderDecoderModel'''), ('''whisper''', '''FlaxWhisperForConditionalGeneration'''), ] ) snake_case_ = OrderedDict( [ ('''whisper''', '''FlaxWhisperForAudioClassification'''), ] ) snake_case_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) snake_case_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) snake_case_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) snake_case_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) snake_case_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : str = FLAX_MODEL_MAPPING snake_case_ = auto_class_update(FlaxAutoModel) class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Dict = FLAX_MODEL_FOR_PRETRAINING_MAPPING snake_case_ = auto_class_update(FlaxAutoModelForPreTraining, head_doc='''pretraining''') class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : List[Any] = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING snake_case_ = auto_class_update(FlaxAutoModelForCausalLM, head_doc='''causal language modeling''') class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Optional[int] = FLAX_MODEL_FOR_MASKED_LM_MAPPING snake_case_ = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='''masked language modeling''') class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Optional[int] = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING snake_case_ = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='''sequence-to-sequence language modeling''', checkpoint_for_example='''t5-base''' ) class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Dict = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING snake_case_ = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='''sequence classification''' ) class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Tuple = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING snake_case_ = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='''question answering''') class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Tuple = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING snake_case_ = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='''token classification''' ) class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : str = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING snake_case_ = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='''multiple choice''') class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Union[str, Any] = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING snake_case_ = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='''next sentence prediction''' ) class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Union[str, Any] = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING snake_case_ = auto_class_update( FlaxAutoModelForImageClassification, head_doc='''image classification''' ) class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : Tuple = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING snake_case_ = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='''vision-to-text modeling''') class SCREAMING_SNAKE_CASE__ (_BaseAutoModelClass ): __lowerCamelCase : List[str] = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING snake_case_ = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='''sequence-to-sequence speech-to-text modeling''' )
216
0
'''simple docstring''' import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def _A ( snake_case , snake_case , snake_case ) -> List[str]: return params[F'''{prefix}/{prefix}/relpos_bias/rel_embedding'''][:, i, :] def _A ( snake_case , snake_case , snake_case , snake_case="attention" ) -> int: _lowercase : Any = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/key/kernel'''][:, i, :, :] ) _lowercase : str = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) _lowercase : int = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/out/kernel'''][:, i, :, :] ) _lowercase : Any = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) _lowercase : List[Any] = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/query/kernel'''][:, i, :, :] ) _lowercase : Any = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) _lowercase : Tuple = np.ascontiguousarray(params[F'''{prefix}/{prefix}/{layer_name}/value/kernel'''][:, i, :, :] ) _lowercase : Tuple = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def _A ( snake_case , snake_case , snake_case , snake_case=False ) -> List[str]: if split_mlp_wi: _lowercase : Any = params[F'''{prefix}/{prefix}/mlp/wi_0/kernel'''][:, i, :] _lowercase : Any = params[F'''{prefix}/{prefix}/mlp/wi_1/kernel'''][:, i, :] _lowercase : List[Any] = (wi_a, wi_a) else: _lowercase : Optional[Any] = params[F'''{prefix}/{prefix}/mlp/wi/kernel'''][:, i, :] _lowercase : Optional[int] = params[F'''{prefix}/{prefix}/mlp/wo/kernel'''][:, i, :] return wi, wo def _A ( snake_case , snake_case , snake_case , snake_case ) -> Union[str, Any]: return params[F'''{prefix}/{prefix}/{layer_name}/scale'''][:, i] def _A ( snake_case , *, snake_case , snake_case , snake_case = False ) -> str: _lowercase : str = traverse_util.flatten_dict(variables["target"] ) _lowercase : Dict = {'''/'''.join(__a ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi _lowercase : str = '''encoder/encoder/mlp/wi_0/kernel''' in old print("Split MLP:" , __a ) _lowercase : Any = collections.OrderedDict() # Shared embeddings. _lowercase : List[str] = old['''token_embedder/embedding'''] # Encoder. for i in range(__a ): # Block i, layer 0 (Self Attention). _lowercase : Any = tax_layer_norm_lookup(__a , __a , "encoder" , "pre_attention_layer_norm" ) _lowercase : List[str] = tax_attention_lookup(__a , __a , "encoder" , "attention" ) _lowercase : str = layer_norm _lowercase : int = k.T _lowercase : Union[str, Any] = o.T _lowercase : Optional[int] = q.T _lowercase : Tuple = v.T # Block i, layer 1 (MLP). _lowercase : Optional[Any] = tax_layer_norm_lookup(__a , __a , "encoder" , "pre_mlp_layer_norm" ) _lowercase : str = tax_mlp_lookup(__a , __a , "encoder" , __a ) _lowercase : Any = layer_norm if split_mlp_wi: _lowercase : List[Any] = wi[0].T _lowercase : Tuple = wi[1].T else: _lowercase : Union[str, Any] = wi.T _lowercase : List[str] = wo.T if scalable_attention: # convert the rel_embedding of each layer _lowercase : List[str] = tax_relpos_bias_lookup( __a , __a , "encoder" ).T _lowercase : str = old['''encoder/encoder_norm/scale'''] if not scalable_attention: _lowercase : Optional[int] = tax_relpos_bias_lookup( __a , 0 , "encoder" ).T _lowercase : int = tax_relpos_bias_lookup( __a , 0 , "decoder" ).T if not is_encoder_only: # Decoder. for i in range(__a ): # Block i, layer 0 (Self Attention). _lowercase : Dict = tax_layer_norm_lookup(__a , __a , "decoder" , "pre_self_attention_layer_norm" ) _lowercase : Optional[int] = tax_attention_lookup(__a , __a , "decoder" , "self_attention" ) _lowercase : List[str] = layer_norm _lowercase : Optional[int] = k.T _lowercase : Union[str, Any] = o.T _lowercase : Optional[int] = q.T _lowercase : str = v.T # Block i, layer 1 (Cross Attention). _lowercase : Union[str, Any] = tax_layer_norm_lookup(__a , __a , "decoder" , "pre_cross_attention_layer_norm" ) _lowercase : Union[str, Any] = tax_attention_lookup(__a , __a , "decoder" , "encoder_decoder_attention" ) _lowercase : Tuple = layer_norm _lowercase : List[str] = k.T _lowercase : List[str] = o.T _lowercase : str = q.T _lowercase : str = v.T # Block i, layer 2 (MLP). _lowercase : Tuple = tax_layer_norm_lookup(__a , __a , "decoder" , "pre_mlp_layer_norm" ) _lowercase : Optional[int] = tax_mlp_lookup(__a , __a , "decoder" , __a ) _lowercase : Dict = layer_norm if split_mlp_wi: _lowercase : List[Any] = wi[0].T _lowercase : Tuple = wi[1].T else: _lowercase : Dict = wi.T _lowercase : Optional[int] = wo.T if scalable_attention: # convert the rel_embedding of each layer _lowercase : Any = tax_relpos_bias_lookup(__a , __a , "decoder" ).T _lowercase : Dict = old['''decoder/decoder_norm/scale'''] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: _lowercase : Tuple = old['''decoder/logits_dense/kernel'''].T return new def _A ( snake_case , snake_case ) -> List[str]: _lowercase : str = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: _lowercase : List[str] = state_dict['''shared.weight'''] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: _lowercase : Tuple = state_dict['''shared.weight'''] if "lm_head.weight" not in state_dict: # For old 1.0 models. print("Using shared word embeddings as lm_head." ) _lowercase : Tuple = state_dict['''shared.weight'''] return state_dict def _A ( snake_case , snake_case , snake_case , snake_case , snake_case ) -> List[Any]: _lowercase : int = checkpoints.load_tax_checkpoint(__a ) _lowercase : List[str] = convert_tax_to_pytorch( __a , num_layers=config.num_layers , is_encoder_only=__a , scalable_attention=__a ) _lowercase : List[str] = make_state_dict(__a , __a ) model.load_state_dict(__a , strict=__a ) def _A ( snake_case , snake_case , snake_case , snake_case = False , snake_case = False , ) -> List[Any]: _lowercase : int = MTaConfig.from_json_file(__a ) print(F'''Building PyTorch model from configuration: {config}''' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: _lowercase : Dict = UMTaEncoderModel(__a ) else: _lowercase : Optional[int] = UMTaForConditionalGeneration(__a ) # Load weights from tf checkpoint load_tax_weights_in_ta(__a , __a , __a , __a , __a ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) model.save_pretrained(__a ) # Verify that we can load the checkpoint. model.from_pretrained(__a ) print("Done" ) if __name__ == "__main__": _snake_case = argparse.ArgumentParser(description='Converts a native T5X checkpoint into a PyTorch checkpoint.') # Required parameters parser.add_argument( '--t5x_checkpoint_path', default=None, type=str, required=True, help='Path to the T5X checkpoint.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--is_encoder_only', action='store_true', help='Check if the model is encoder-decoder model', default=False ) parser.add_argument( '--scalable_attention', action='store_true', help='Whether the model uses scaled attention (umt5 model)', default=False, ) _snake_case = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
250
'''simple docstring''' from __future__ import annotations from fractions import Fraction from math import gcd, sqrt def a ( __a ) -> bool: '''simple docstring''' UpperCamelCase__ :int = int(number**0.5 ) return number == sq * sq def a ( __a , __a , __a , __a , __a , __a ) -> tuple[int, int]: '''simple docstring''' UpperCamelCase__ :int = x_num * y_den * z_den + y_num * x_den * z_den + z_num * x_den * y_den UpperCamelCase__ :int = x_den * y_den * z_den UpperCamelCase__ :int = gcd(__a , __a ) top //= hcf bottom //= hcf return top, bottom def a ( __a = 35 ) -> int: '''simple docstring''' UpperCamelCase__ :set = set() UpperCamelCase__ :int UpperCamelCase__ :Fraction = Fraction(0 ) UpperCamelCase__ :tuple[int, int] for x_num in range(1 , order + 1 ): for x_den in range(x_num + 1 , order + 1 ): for y_num in range(1 , order + 1 ): for y_den in range(y_num + 1 , order + 1 ): # n=1 UpperCamelCase__ :int = x_num * y_den + x_den * y_num UpperCamelCase__ :Any = x_den * y_den UpperCamelCase__ :Tuple = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCamelCase__ :Tuple = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) # n=2 UpperCamelCase__ :List[str] = ( x_num * x_num * y_den * y_den + x_den * x_den * y_num * y_num ) UpperCamelCase__ :Dict = x_den * x_den * y_den * y_den if is_sq(__a ) and is_sq(__a ): UpperCamelCase__ :Any = int(sqrt(__a ) ) UpperCamelCase__ :Optional[int] = int(sqrt(__a ) ) UpperCamelCase__ :int = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCamelCase__ :Tuple = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) # n=-1 UpperCamelCase__ :Tuple = x_num * y_num UpperCamelCase__ :Union[str, Any] = x_den * y_num + x_num * y_den UpperCamelCase__ :List[str] = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCamelCase__ :Union[str, Any] = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) # n=2 UpperCamelCase__ :Optional[Any] = x_num * x_num * y_num * y_num UpperCamelCase__ :Tuple = ( x_den * x_den * y_num * y_num + x_num * x_num * y_den * y_den ) if is_sq(__a ) and is_sq(__a ): UpperCamelCase__ :str = int(sqrt(__a ) ) UpperCamelCase__ :Any = int(sqrt(__a ) ) UpperCamelCase__ :Dict = gcd(__a , __a ) z_num //= hcf z_den //= hcf if 0 < z_num < z_den <= order: UpperCamelCase__ :int = add_three( __a , __a , __a , __a , __a , __a ) unique_s.add(__a ) for num, den in unique_s: total += Fraction(__a , __a ) return total.denominator + total.numerator if __name__ == "__main__": print(F"""{solution() = }""")
97
0
"""simple docstring""" import argparse import torch # Step 1. clone https://github.com/microsoft/unilm # Step 2. git checkout to https://github.com/microsoft/unilm/commit/b94ec76c36f02fb2b0bf0dcb0b8554a2185173cd # Step 3. cd unilm # Step 4. ln -s $(realpath wavlm/modules.py) ./ # create simlink # import classes from unilm.wavlm.WavLM import WavLM as WavLMOrig from unilm.wavlm.WavLM import WavLMConfig as WavLMConfigOrig from transformers import WavLMConfig, WavLMModel, logging logging.set_verbosity_info() _lowercase : str = logging.get_logger(__name__) _lowercase : Tuple = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn.grep_linear': 'encoder.layers.*.attention.gru_rel_pos_linear', 'self_attn.relative_attention_bias': 'encoder.layers.*.attention.rel_attn_embed', 'self_attn.grep_a': 'encoder.layers.*.attention.gru_rel_pos_const', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'ctc_proj', 'mask_emb': 'masked_spec_embed', } _lowercase : Optional[Any] = [ 'ctc_proj', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def lowercase__ ( snake_case_ :Any , snake_case_ :Optional[Any] , snake_case_ :Optional[Any] , snake_case_ :Optional[int] , snake_case_ :Optional[int] ): for attribute in key.split('''.''' ): __UpperCAmelCase = getattr(snake_case_ , snake_case_ ) if weight_type is not None: __UpperCAmelCase = getattr(snake_case_ , snake_case_ ).shape else: __UpperCAmelCase = hf_pointer.shape assert hf_shape == value.shape, ( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''' ) if weight_type == "weight": __UpperCAmelCase = value elif weight_type == "weight_g": __UpperCAmelCase = value elif weight_type == "weight_v": __UpperCAmelCase = value elif weight_type == "bias": __UpperCAmelCase = value else: __UpperCAmelCase = value logger.info(F'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def lowercase__ ( snake_case_ :Union[str, Any] , snake_case_ :Dict ): __UpperCAmelCase = [] __UpperCAmelCase = fairseq_model.state_dict() __UpperCAmelCase = hf_model.feature_extractor for name, value in fairseq_dict.items(): __UpperCAmelCase = False if "conv_layers" in name: load_conv_layer( snake_case_ , snake_case_ , snake_case_ , snake_case_ , hf_model.config.feat_extract_norm == '''group''' , ) __UpperCAmelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __UpperCAmelCase = True if "*" in mapped_key: __UpperCAmelCase = name.split(snake_case_ )[0].split('''.''' )[-2] __UpperCAmelCase = mapped_key.replace('''*''' , snake_case_ ) if "weight_g" in name: __UpperCAmelCase = '''weight_g''' elif "weight_v" in name: __UpperCAmelCase = '''weight_v''' elif "bias" in name and "relative_attention_bias" not in name: __UpperCAmelCase = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __UpperCAmelCase = '''weight''' else: __UpperCAmelCase = None set_recursively(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) continue if not is_used: unused_weights.append(snake_case_ ) logger.warning(F'''Unused weights: {unused_weights}''' ) def lowercase__ ( snake_case_ :Any , snake_case_ :Union[str, Any] , snake_case_ :Dict , snake_case_ :Optional[int] , snake_case_ :Optional[int] ): __UpperCAmelCase = full_name.split('''conv_layers.''' )[-1] __UpperCAmelCase = name.split('''.''' ) __UpperCAmelCase = int(items[0] ) __UpperCAmelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) __UpperCAmelCase = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) __UpperCAmelCase = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) __UpperCAmelCase = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) __UpperCAmelCase = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(snake_case_ ) @torch.no_grad() def lowercase__ ( snake_case_ :str , snake_case_ :int , snake_case_ :Any=None ): # load the pre-trained checkpoints __UpperCAmelCase = torch.load(snake_case_ ) __UpperCAmelCase = WavLMConfigOrig(checkpoint['''cfg'''] ) __UpperCAmelCase = WavLMOrig(snake_case_ ) model.load_state_dict(checkpoint['''model'''] ) model.eval() if config_path is not None: __UpperCAmelCase = WavLMConfig.from_pretrained(snake_case_ ) else: __UpperCAmelCase = WavLMConfig() __UpperCAmelCase = WavLMModel(snake_case_ ) recursively_load_weights(snake_case_ , snake_case_ ) hf_wavlm.save_pretrained(snake_case_ ) if __name__ == "__main__": _lowercase : List[str] = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') _lowercase : Optional[Any] = parser.parse_args() convert_wavlm_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
358
"""simple docstring""" import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class _UpperCAmelCase : def __init__( self : Optional[int] , _lowercase : Any , _lowercase : List[str]=14 , _lowercase : Dict=7 , _lowercase : Optional[int]=True , _lowercase : Optional[int]=True , _lowercase : Any=False , _lowercase : Any=True , _lowercase : List[str]=99 , _lowercase : int=32 , _lowercase : Union[str, Any]=4 , _lowercase : Dict=4 , _lowercase : List[Any]=4 , _lowercase : Dict=37 , _lowercase : Tuple="gelu" , _lowercase : Optional[int]=0.1 , _lowercase : Dict=0.1 , _lowercase : Union[str, Any]=5_12 , _lowercase : int=0.02 , ): __UpperCAmelCase = parent __UpperCAmelCase = batch_size __UpperCAmelCase = seq_length __UpperCAmelCase = is_training __UpperCAmelCase = use_input_mask __UpperCAmelCase = use_token_type_ids __UpperCAmelCase = use_labels __UpperCAmelCase = vocab_size __UpperCAmelCase = hidden_size __UpperCAmelCase = rotary_dim __UpperCAmelCase = num_hidden_layers __UpperCAmelCase = num_attention_heads __UpperCAmelCase = intermediate_size __UpperCAmelCase = hidden_act __UpperCAmelCase = hidden_dropout_prob __UpperCAmelCase = attention_probs_dropout_prob __UpperCAmelCase = max_position_embeddings __UpperCAmelCase = initializer_range __UpperCAmelCase = None __UpperCAmelCase = vocab_size - 1 __UpperCAmelCase = vocab_size - 1 __UpperCAmelCase = vocab_size - 1 def a ( self : int ): __UpperCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase = None if self.use_input_mask: __UpperCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=_lowercase , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def a ( self : str ): __UpperCAmelCase = self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = config_and_inputs __UpperCAmelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict def a ( self : List[Any] , _lowercase : Tuple , _lowercase : Optional[Any] , _lowercase : List[str] , _lowercase : List[str] ): __UpperCAmelCase = 20 __UpperCAmelCase = model_class_name(_lowercase ) __UpperCAmelCase = model.init_cache(input_ids.shape[0] , _lowercase ) __UpperCAmelCase = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype='''i4''' ) __UpperCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __UpperCAmelCase = model( input_ids[:, :-1] , attention_mask=_lowercase , past_key_values=_lowercase , position_ids=_lowercase , ) __UpperCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='''i4''' ) __UpperCAmelCase = model( input_ids[:, -1:] , attention_mask=_lowercase , past_key_values=outputs_cache.past_key_values , position_ids=_lowercase , ) __UpperCAmelCase = model(_lowercase ) __UpperCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) def a ( self : List[Any] , _lowercase : Optional[int] , _lowercase : Any , _lowercase : Optional[int] , _lowercase : Union[str, Any] ): __UpperCAmelCase = 20 __UpperCAmelCase = model_class_name(_lowercase ) __UpperCAmelCase = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) __UpperCAmelCase = model.init_cache(input_ids.shape[0] , _lowercase ) __UpperCAmelCase = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __UpperCAmelCase = model( input_ids[:, :-1] , attention_mask=_lowercase , past_key_values=_lowercase , position_ids=_lowercase , ) __UpperCAmelCase = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='''i4''' ) __UpperCAmelCase = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=_lowercase , position_ids=_lowercase , ) __UpperCAmelCase = model(_lowercase , attention_mask=_lowercase ) __UpperCAmelCase = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=F'''Max diff is {diff}''' ) @require_flax class _UpperCAmelCase ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): a__ : Any = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () a__ : List[Any] = (FlaxGPTJForCausalLM,) if is_flax_available() else () def a ( self : List[Any] ): __UpperCAmelCase = FlaxGPTJModelTester(self ) def a ( self : Any ): for model_class_name in self.all_model_classes: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(_lowercase , _lowercase , _lowercase , _lowercase ) def a ( self : Union[str, Any] ): for model_class_name in self.all_model_classes: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( _lowercase , _lowercase , _lowercase , _lowercase ) @tooslow def a ( self : Tuple ): __UpperCAmelCase = GPTaTokenizer.from_pretrained('''gpt2''' , pad_token='''<|endoftext|>''' , padding_side='''left''' ) __UpperCAmelCase = tokenizer(['''Hello this is a long string''', '''Hey'''] , return_tensors='''np''' , padding=_lowercase , truncation=_lowercase ) __UpperCAmelCase = FlaxGPTJForCausalLM.from_pretrained('''EleutherAI/gpt-j-6B''' ) __UpperCAmelCase = False __UpperCAmelCase = model.config.eos_token_id __UpperCAmelCase = jax.jit(model.generate ) __UpperCAmelCase = jit_generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , pad_token_id=tokenizer.pad_token_id ).sequences __UpperCAmelCase = tokenizer.batch_decode(_lowercase , skip_special_tokens=_lowercase ) __UpperCAmelCase = [ '''Hello this is a long string of text.\n\nI\'m trying to get the text of the''', '''Hey, I\'m a little late to the party. I\'m going to''', ] self.assertListEqual(_lowercase , _lowercase ) @is_pt_flax_cross_test def a ( self : Tuple ): __UpperCAmelCase , __UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __UpperCAmelCase = self._prepare_for_class(_lowercase , _lowercase ) __UpperCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __UpperCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __UpperCAmelCase = getattr(_lowercase , _lowercase ) __UpperCAmelCase , __UpperCAmelCase = pt_inputs['''input_ids'''].shape __UpperCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_lowercase ): __UpperCAmelCase = 0 __UpperCAmelCase = 1 __UpperCAmelCase = 0 __UpperCAmelCase = 1 __UpperCAmelCase = pt_model_class(_lowercase ).eval() __UpperCAmelCase = model_class(_lowercase , dtype=jnp.floataa ) __UpperCAmelCase = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _lowercase ) __UpperCAmelCase = fx_state with torch.no_grad(): __UpperCAmelCase = pt_model(**_lowercase ).to_tuple() __UpperCAmelCase = fx_model(**_lowercase ).to_tuple() self.assertEqual(len(_lowercase ) , len(_lowercase ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(_lowercase , _lowercase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_lowercase ) __UpperCAmelCase = model_class.from_pretrained(_lowercase , from_pt=_lowercase ) __UpperCAmelCase = fx_model_loaded(**_lowercase ).to_tuple() self.assertEqual( len(_lowercase ) , len(_lowercase ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output_loaded, pt_output in zip(_lowercase , _lowercase ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def a ( self : Any ): __UpperCAmelCase , __UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __UpperCAmelCase = self._prepare_for_class(_lowercase , _lowercase ) __UpperCAmelCase = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __UpperCAmelCase = model_class.__name__[4:] # Skip the "Flax" at the beginning __UpperCAmelCase = getattr(_lowercase , _lowercase ) __UpperCAmelCase = pt_model_class(_lowercase ).eval() __UpperCAmelCase = model_class(_lowercase , dtype=jnp.floataa ) __UpperCAmelCase = load_flax_weights_in_pytorch_model(_lowercase , fx_model.params ) __UpperCAmelCase , __UpperCAmelCase = pt_inputs['''input_ids'''].shape __UpperCAmelCase = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_lowercase ): __UpperCAmelCase = 0 __UpperCAmelCase = 1 __UpperCAmelCase = 0 __UpperCAmelCase = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): __UpperCAmelCase = pt_model(**_lowercase ).to_tuple() __UpperCAmelCase = fx_model(**_lowercase ).to_tuple() self.assertEqual(len(_lowercase ) , len(_lowercase ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(_lowercase , _lowercase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_lowercase ) __UpperCAmelCase = pt_model_class.from_pretrained(_lowercase , from_flax=_lowercase ) with torch.no_grad(): __UpperCAmelCase = pt_model_loaded(**_lowercase ).to_tuple() self.assertEqual( len(_lowercase ) , len(_lowercase ) , '''Output lengths differ between Flax and PyTorch''' ) for fx_output, pt_output in zip(_lowercase , _lowercase ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def a ( self : Tuple ): for model_class_name in self.all_model_classes: __UpperCAmelCase = model_class_name.from_pretrained('''EleutherAI/gpt-j-6B''' ) __UpperCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(_lowercase )
86
0
from typing import Any class A_ : def __init__( self : Optional[int] ,SCREAMING_SNAKE_CASE__ : Any): __lowerCamelCase : Union[str, Any] = data __lowerCamelCase : int = None class A_ : def __init__( self : List[Any]): __lowerCamelCase : Optional[Any] = None def lowerCAmelCase ( self : Any): __lowerCamelCase : Tuple = self.head while temp is not None: print(temp.data ,end=' ') __lowerCamelCase : Optional[Any] = temp.next print() def lowerCAmelCase ( self : str ,SCREAMING_SNAKE_CASE__ : Any): __lowerCamelCase : Optional[int] = Node(SCREAMING_SNAKE_CASE__) __lowerCamelCase : Tuple = self.head __lowerCamelCase : Optional[Any] = new_node def lowerCAmelCase ( self : Dict ,SCREAMING_SNAKE_CASE__ : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : Optional[Any]): if node_data_a == node_data_a: return else: __lowerCamelCase : Union[str, Any] = self.head while node_a is not None and node_a.data != node_data_a: __lowerCamelCase : Optional[Any] = node_a.next __lowerCamelCase : int = self.head while node_a is not None and node_a.data != node_data_a: __lowerCamelCase : Tuple = node_a.next if node_a is None or node_a is None: return __lowerCamelCase , __lowerCamelCase : str = node_a.data, node_a.data if __name__ == "__main__": a =LinkedList() for i in range(5, 0, -1): ll.push(i) ll.print_list() ll.swap_nodes(1, 4) print("""After swapping""") ll.print_list()
73
from math import factorial def lowerCamelCase__ (_UpperCAmelCase = 100): return sum(int(_UpperCAmelCase) for x in str(factorial(_UpperCAmelCase))) if __name__ == "__main__": print(solution(int(input('Enter the Number: ').strip())))
137
0
'''simple docstring''' def _A ( A__ = 1000 ): """simple docstring""" __lowercase , __lowercase = 1, 1 __lowercase = [] for i in range(1 , n + 1 ): __lowercase = prev_numerator + 2 * prev_denominator __lowercase = prev_numerator + prev_denominator if len(str(A__ ) ) > len(str(A__ ) ): result.append(A__ ) __lowercase = numerator __lowercase = denominator return len(A__ ) if __name__ == "__main__": print(f'{solution() = }')
52
'''simple docstring''' lowerCAmelCase__ = { "joule": 1.0, "kilojoule": 1000, "megajoule": 100_0000, "gigajoule": 10_0000_0000, "wattsecond": 1.0, "watthour": 3600, "kilowatthour": 360_0000, "newtonmeter": 1.0, "calorie_nutr": 4186.8, "kilocalorie_nutr": 418_6800.00, "electronvolt": 1.602_176_634e-19, "britishthermalunit_it": 1055.0_5585, "footpound": 1.355_818, } def _A ( A__ , A__ , A__ ): """simple docstring""" if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: __lowercase = ( F"Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n" F"Valid values are: {', '.join(A__ )}" ) raise ValueError(A__ ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
52
1
"""simple docstring""" import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class a__ : def lowercase ( self : int ) -> Any: torch.manual_seed(0 ) lowercase : List[str] = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) lowercase : str = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) lowercase : Optional[int] = UNetaDConditionModel( sample_size=32, layers_per_block=1, block_out_channels=[32, 64], down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ], mid_block_type='UNetMidBlock2DSimpleCrossAttn', up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'], in_channels=3, out_channels=6, cross_attention_dim=32, encoder_hid_dim=32, attention_head_dim=8, addition_embed_type='text', addition_embed_type_num_heads=2, cross_attention_norm='group_norm', resnet_time_scale_shift='scale_shift', act_fn='gelu', ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase : Union[str, Any] = DDPMScheduler( num_train_timesteps=1000, beta_schedule='squaredcos_cap_v2', beta_start=0.0001, beta_end=0.02, thresholding=lowerCAmelCase, dynamic_thresholding_ratio=0.95, sample_max_value=1.0, prediction_type='epsilon', variance_type='learned_range', ) torch.manual_seed(0 ) lowercase : str = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def lowercase ( self : Optional[int] ) -> Optional[Any]: torch.manual_seed(0 ) lowercase : str = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) lowercase : Tuple = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) lowercase : Dict = UNetaDConditionModel( sample_size=32, layers_per_block=[1, 2], block_out_channels=[32, 64], down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ], mid_block_type='UNetMidBlock2DSimpleCrossAttn', up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'], in_channels=6, out_channels=6, cross_attention_dim=32, encoder_hid_dim=32, attention_head_dim=8, addition_embed_type='text', addition_embed_type_num_heads=2, cross_attention_norm='group_norm', resnet_time_scale_shift='scale_shift', act_fn='gelu', class_embed_type='timestep', mid_block_scale_factor=1.414, time_embedding_act_fn='gelu', time_embedding_dim=32, ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) lowercase : Any = DDPMScheduler( num_train_timesteps=1000, beta_schedule='squaredcos_cap_v2', beta_start=0.0001, beta_end=0.02, thresholding=lowerCAmelCase, dynamic_thresholding_ratio=0.95, sample_max_value=1.0, prediction_type='epsilon', variance_type='learned_range', ) torch.manual_seed(0 ) lowercase : Optional[int] = DDPMScheduler( num_train_timesteps=1000, beta_schedule='squaredcos_cap_v2', beta_start=0.0001, beta_end=0.02, ) torch.manual_seed(0 ) lowercase : str = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def lowercase ( self : Tuple ) -> Optional[Any]: lowercase : Optional[Any] = self.get_dummy_components() lowercase : Optional[int] = self.pipeline_class(**lowerCAmelCase ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowercase : Optional[Any] = self.get_dummy_inputs(lowerCAmelCase ) lowercase : Union[str, Any] = inputs['prompt'] lowercase : List[Any] = inputs['generator'] lowercase : Union[str, Any] = inputs['num_inference_steps'] lowercase : Optional[int] = inputs['output_type'] if "image" in inputs: lowercase : str = inputs['image'] else: lowercase : str = None if "mask_image" in inputs: lowercase : str = inputs['mask_image'] else: lowercase : Any = None if "original_image" in inputs: lowercase : Optional[int] = inputs['original_image'] else: lowercase : Optional[Any] = None lowercase , lowercase : Union[str, Any] = pipe.encode_prompt(lowerCAmelCase ) # inputs with prompt converted to embeddings lowercase : int = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: lowercase : int = image if mask_image is not None: lowercase : Dict = mask_image if original_image is not None: lowercase : Optional[int] = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(lowerCAmelCase, lowerCAmelCase, lowerCAmelCase ) lowercase : int = pipe(**lowerCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCAmelCase ) lowercase : Union[str, Any] = self.pipeline_class.from_pretrained(lowerCAmelCase ) pipe_loaded.to(lowerCAmelCase ) pipe_loaded.set_progress_bar_config(disable=lowerCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(lowerCAmelCase, lowerCAmelCase ) is None, f'''`{optional_component}` did not stay set to None after loading.''', ) lowercase : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase ) lowercase : str = inputs['generator'] lowercase : Union[str, Any] = inputs['num_inference_steps'] lowercase : Dict = inputs['output_type'] # inputs with prompt converted to embeddings lowercase : Optional[Any] = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: lowercase : str = image if mask_image is not None: lowercase : Dict = mask_image if original_image is not None: lowercase : Optional[Any] = original_image lowercase : str = pipe_loaded(**lowerCAmelCase )[0] lowercase : List[str] = np.abs(to_np(lowerCAmelCase ) - to_np(lowerCAmelCase ) ).max() self.assertLess(lowerCAmelCase, 1e-4 ) def lowercase ( self : str ) -> Dict: lowercase : Dict = self.get_dummy_components() lowercase : Optional[Any] = self.pipeline_class(**lowerCAmelCase ) pipe.to(lowerCAmelCase ) pipe.set_progress_bar_config(disable=lowerCAmelCase ) lowercase : Optional[int] = self.get_dummy_inputs(lowerCAmelCase ) lowercase : Tuple = pipe(**lowerCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(lowerCAmelCase ) lowercase : Union[str, Any] = self.pipeline_class.from_pretrained(lowerCAmelCase ) pipe_loaded.to(lowerCAmelCase ) pipe_loaded.set_progress_bar_config(disable=lowerCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests lowercase : str = self.get_dummy_inputs(lowerCAmelCase ) lowercase : List[Any] = pipe_loaded(**lowerCAmelCase )[0] lowercase : int = np.abs(to_np(lowerCAmelCase ) - to_np(lowerCAmelCase ) ).max() self.assertLess(lowerCAmelCase, 1e-4 )
255
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _UpperCamelCase: List[str] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase: Union[str, Any] = ['NllbTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase: Optional[int] = ['NllbTokenizerFast'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys _UpperCamelCase: Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
255
1
'''simple docstring''' from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name _lowerCAmelCase = "\n Examples:\n ```py\n >>> import torch\n >>> import numpy as np\n\n >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline\n >>> from transformers import pipeline\n >>> from diffusers.utils import load_image\n\n\n >>> def make_hint(image, depth_estimator):\n ... image = depth_estimator(image)[\"depth\"]\n ... image = np.array(image)\n ... image = image[:, :, None]\n ... image = np.concatenate([image, image, image], axis=2)\n ... detected_map = torch.from_numpy(image).float() / 255.0\n ... hint = detected_map.permute(2, 0, 1)\n ... return hint\n\n\n >>> depth_estimator = pipeline(\"depth-estimation\")\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... \"kandinsky-community/kandinsky-2-2-prior\", torch_dtype=torch.float16\n ... )\n >>> pipe_prior = pipe_prior.to(\"cuda\")\n\n >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained(\n ... \"kandinsky-community/kandinsky-2-2-controlnet-depth\", torch_dtype=torch.float16\n ... )\n >>> pipe = pipe.to(\"cuda\")\n\n\n >>> img = load_image(\n ... \"https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main\"\n ... \"/kandinsky/cat.png\"\n ... ).resize((768, 768))\n\n >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to(\"cuda\")\n\n >>> prompt = \"A robot, 4k photo\"\n >>> negative_prior_prompt = \"lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature\"\n\n >>> generator = torch.Generator(device=\"cuda\").manual_seed(43)\n\n >>> image_emb, zero_image_emb = pipe_prior(\n ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator\n ... ).to_tuple()\n\n >>> images = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... hint=hint,\n ... num_inference_steps=50,\n ... generator=generator,\n ... height=768,\n ... width=768,\n ... ).images\n\n >>> images[0].save(\"robot_cat.png\")\n ```\n" def UpperCamelCase ( a , a , a=8 ) -> Any: '''simple docstring''' __magic_name__ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __magic_name__ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class _SCREAMING_SNAKE_CASE ( __a ): def __init__( self : Union[str, Any] , a__ : UNetaDConditionModel , a__ : DDPMScheduler , a__ : VQModel , ): super().__init__() self.register_modules( unet=a__ , scheduler=a__ , movq=a__ , ) __magic_name__ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def snake_case__ ( self : Union[str, Any] , a__ : List[Any] , a__ : List[str] , a__ : Tuple , a__ : int , a__ : Tuple , a__ : Tuple ): if latents is None: __magic_name__ = randn_tensor(a__ , generator=a__ , device=a__ , dtype=a__ ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) __magic_name__ = latents.to(a__ ) __magic_name__ = latents * scheduler.init_noise_sigma return latents def snake_case__ ( self : int , a__ : Optional[Any]=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) __magic_name__ = torch.device(F'''cuda:{gpu_id}''' ) __magic_name__ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(a__ , a__ ) def snake_case__ ( self : Dict , a__ : List[Any]=0 ): if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) __magic_name__ = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=a__ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __magic_name__ = None for cpu_offloaded_model in [self.unet, self.movq]: __magic_name__ , __magic_name__ = cpu_offload_with_hook(a__ , a__ , prev_module_hook=a__ ) # We'll offload the last model manually. __magic_name__ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def snake_case__ ( self : Tuple ): if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(a__ , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(a__ ) def __call__( self : int , a__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , a__ : Union[torch.FloatTensor, List[torch.FloatTensor]] , a__ : torch.FloatTensor , a__ : int = 512 , a__ : int = 512 , a__ : int = 100 , a__ : float = 4.0 , a__ : int = 1 , a__ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a__ : Optional[torch.FloatTensor] = None , a__ : Optional[str] = "pil" , a__ : bool = True , ): __magic_name__ = self._execution_device __magic_name__ = guidance_scale > 1.0 if isinstance(a__ , a__ ): __magic_name__ = torch.cat(a__ , dim=0 ) if isinstance(a__ , a__ ): __magic_name__ = torch.cat(a__ , dim=0 ) if isinstance(a__ , a__ ): __magic_name__ = torch.cat(a__ , dim=0 ) __magic_name__ = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: __magic_name__ = image_embeds.repeat_interleave(a__ , dim=0 ) __magic_name__ = negative_image_embeds.repeat_interleave(a__ , dim=0 ) __magic_name__ = hint.repeat_interleave(a__ , dim=0 ) __magic_name__ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=a__ ) __magic_name__ = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=a__ ) self.scheduler.set_timesteps(a__ , device=a__ ) __magic_name__ = self.scheduler.timesteps __magic_name__ = self.movq.config.latent_channels __magic_name__ , __magic_name__ = downscale_height_and_width(a__ , a__ , self.movq_scale_factor ) # create initial latent __magic_name__ = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , a__ , a__ , a__ , self.scheduler , ) for i, t in enumerate(self.progress_bar(a__ ) ): # expand the latents if we are doing classifier free guidance __magic_name__ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __magic_name__ = {'''image_embeds''': image_embeds, '''hint''': hint} __magic_name__ = self.unet( sample=a__ , timestep=a__ , encoder_hidden_states=a__ , added_cond_kwargs=a__ , return_dict=a__ , )[0] if do_classifier_free_guidance: __magic_name__ , __magic_name__ = noise_pred.split(latents.shape[1] , dim=1 ) __magic_name__ , __magic_name__ = noise_pred.chunk(2 ) __magic_name__ , __magic_name__ = variance_pred.chunk(2 ) __magic_name__ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __magic_name__ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __magic_name__ , __magic_name__ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __magic_name__ = self.scheduler.step( a__ , a__ , a__ , generator=a__ , )[0] # post-processing __magic_name__ = self.movq.decode(a__ , force_not_quantize=a__ )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: __magic_name__ = image * 0.5 + 0.5 __magic_name__ = image.clamp(0 , 1 ) __magic_name__ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __magic_name__ = self.numpy_to_pil(a__ ) if not return_dict: return (image,) return ImagePipelineOutput(images=a__ )
98
'''simple docstring''' import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() _lowerCAmelCase = logging.get_logger(__name__) def UpperCamelCase ( a ) -> str: '''simple docstring''' __magic_name__ = SwinConfig.from_pretrained( '''microsoft/swin-tiny-patch4-window7-224''' , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) __magic_name__ = MaskFormerConfig(backbone_config=a ) __magic_name__ = '''huggingface/label-files''' if "ade20k-full" in model_name: # this should be ok __magic_name__ = 847 __magic_name__ = '''maskformer-ade20k-full-id2label.json''' elif "ade" in model_name: # this should be ok __magic_name__ = 150 __magic_name__ = '''ade20k-id2label.json''' elif "coco-stuff" in model_name: # this should be ok __magic_name__ = 171 __magic_name__ = '''maskformer-coco-stuff-id2label.json''' elif "coco" in model_name: # TODO __magic_name__ = 133 __magic_name__ = '''coco-panoptic-id2label.json''' elif "cityscapes" in model_name: # this should be ok __magic_name__ = 19 __magic_name__ = '''cityscapes-id2label.json''' elif "vistas" in model_name: # this should be ok __magic_name__ = 65 __magic_name__ = '''mapillary-vistas-id2label.json''' __magic_name__ = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) __magic_name__ = {int(a ): v for k, v in idalabel.items()} return config def UpperCamelCase ( a ) -> Tuple: '''simple docstring''' __magic_name__ = [] # stem # fmt: off rename_keys.append(('''backbone.patch_embed.proj.weight''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.proj.bias''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''model.pixel_level_module.encoder.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''model.pixel_level_module.encoder.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm1.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm1.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.relative_position_index''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.proj.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.attn.proj.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm2.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.norm2.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc1.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc1.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc2.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((F'''backbone.layers.{i}.blocks.{j}.mlp.fc2.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((F'''backbone.layers.{i}.downsample.reduction.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((F'''backbone.layers.{i}.downsample.norm.weight''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((F'''backbone.layers.{i}.downsample.norm.bias''', F'''model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((F'''backbone.norm{i}.weight''', F'''model.pixel_level_module.encoder.hidden_states_norms.{i}.weight''') ) rename_keys.append((F'''backbone.norm{i}.bias''', F'''model.pixel_level_module.encoder.hidden_states_norms.{i}.bias''') ) # FPN rename_keys.append(('''sem_seg_head.layer_4.weight''', '''model.pixel_level_module.decoder.fpn.stem.0.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.weight''', '''model.pixel_level_module.decoder.fpn.stem.1.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.bias''', '''model.pixel_level_module.decoder.fpn.stem.1.bias''') ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((F'''sem_seg_head.adapter_{source_index}.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight''') ) rename_keys.append((F'''sem_seg_head.adapter_{source_index}.norm.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight''') ) rename_keys.append((F'''sem_seg_head.adapter_{source_index}.norm.bias''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.norm.weight''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight''') ) rename_keys.append((F'''sem_seg_head.layer_{source_index}.norm.bias''', F'''model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias''') ) rename_keys.append(('''sem_seg_head.mask_features.weight''', '''model.pixel_level_module.decoder.mask_projection.weight''') ) rename_keys.append(('''sem_seg_head.mask_features.bias''', '''model.pixel_level_module.decoder.mask_projection.bias''') ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight''', F'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias''', F'''model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias''') ) # cross-attention out projection rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias''') ) # MLP 1 rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight''', F'''model.transformer_module.decoder.layers.{idx}.fc1.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias''', F'''model.transformer_module.decoder.layers.{idx}.fc1.bias''') ) # MLP 2 rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight''', F'''model.transformer_module.decoder.layers.{idx}.fc2.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias''', F'''model.transformer_module.decoder.layers.{idx}.fc2.bias''') ) # layernorm 1 (self-attention layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight''', F'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias''', F'''model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias''') ) # layernorm 2 (cross-attention layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias''', F'''model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias''') ) # layernorm 3 (final layernorm) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight''', F'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias''', F'''model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.weight''', '''model.transformer_module.decoder.layernorm.weight''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.bias''', '''model.transformer_module.decoder.layernorm.bias''') ) # heads on top rename_keys.append(('''sem_seg_head.predictor.query_embed.weight''', '''model.transformer_module.queries_embedder.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.weight''', '''model.transformer_module.input_projection.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.bias''', '''model.transformer_module.input_projection.bias''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.weight''', '''class_predictor.weight''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.bias''', '''class_predictor.bias''') ) for i in range(3 ): rename_keys.append((F'''sem_seg_head.predictor.mask_embed.layers.{i}.weight''', F'''mask_embedder.{i}.0.weight''') ) rename_keys.append((F'''sem_seg_head.predictor.mask_embed.layers.{i}.bias''', F'''mask_embedder.{i}.0.bias''') ) # fmt: on return rename_keys def UpperCamelCase ( a , a , a ) -> str: '''simple docstring''' __magic_name__ = dct.pop(a ) __magic_name__ = val def UpperCamelCase ( a , a ) -> List[str]: '''simple docstring''' __magic_name__ = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __magic_name__ = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __magic_name__ = state_dict.pop(F'''backbone.layers.{i}.blocks.{j}.attn.qkv.weight''' ) __magic_name__ = state_dict.pop(F'''backbone.layers.{i}.blocks.{j}.attn.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ = in_proj_weight[:dim, :] __magic_name__ = in_proj_bias[: dim] __magic_name__ = in_proj_weight[ dim : dim * 2, : ] __magic_name__ = in_proj_bias[ dim : dim * 2 ] __magic_name__ = in_proj_weight[ -dim :, : ] __magic_name__ = in_proj_bias[-dim :] # fmt: on def UpperCamelCase ( a , a ) -> int: '''simple docstring''' # fmt: off __magic_name__ = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight''' ) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ = in_proj_weight[: hidden_size, :] __magic_name__ = in_proj_bias[:config.hidden_size] __magic_name__ = in_proj_weight[hidden_size : hidden_size * 2, :] __magic_name__ = in_proj_bias[hidden_size : hidden_size * 2] __magic_name__ = in_proj_weight[-hidden_size :, :] __magic_name__ = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight''' ) __magic_name__ = state_dict.pop(F'''sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias''' ) # next, add query, keys and values (in that order) to the state dict __magic_name__ = in_proj_weight[: hidden_size, :] __magic_name__ = in_proj_bias[:config.hidden_size] __magic_name__ = in_proj_weight[hidden_size : hidden_size * 2, :] __magic_name__ = in_proj_bias[hidden_size : hidden_size * 2] __magic_name__ = in_proj_weight[-hidden_size :, :] __magic_name__ = in_proj_bias[-hidden_size :] # fmt: on def UpperCamelCase ( ) -> torch.Tensor: '''simple docstring''' __magic_name__ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __magic_name__ = Image.open(requests.get(a , stream=a ).raw ) return im @torch.no_grad() def UpperCamelCase ( a , a , a , a = False ) -> Dict: '''simple docstring''' __magic_name__ = get_maskformer_config(a ) # load original state_dict with open(a , '''rb''' ) as f: __magic_name__ = pickle.load(a ) __magic_name__ = data['''model'''] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys __magic_name__ = create_rename_keys(a ) for src, dest in rename_keys: rename_key(a , a , a ) read_in_swin_q_k_v(a , config.backbone_config ) read_in_decoder_q_k_v(a , a ) # update to torch tensors for key, value in state_dict.items(): __magic_name__ = torch.from_numpy(a ) # load 🤗 model __magic_name__ = MaskFormerForInstanceSegmentation(a ) model.eval() for name, param in model.named_parameters(): print(a , param.shape ) __magic_name__ , __magic_name__ = model.load_state_dict(a , strict=a ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(a ) == 0, F'''Unexpected keys: {unexpected_keys}''' # verify results __magic_name__ = prepare_img() if "vistas" in model_name: __magic_name__ = 65 elif "cityscapes" in model_name: __magic_name__ = 6_5535 else: __magic_name__ = 255 __magic_name__ = True if '''ade''' in model_name else False __magic_name__ = MaskFormerImageProcessor(ignore_index=a , reduce_labels=a ) __magic_name__ = image_processor(a , return_tensors='''pt''' ) __magic_name__ = model(**a ) print('''Logits:''' , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": __magic_name__ = torch.tensor( [[3.63_53, -4.47_70, -2.60_65], [0.50_81, -4.23_94, -3.53_43], [2.19_09, -5.03_53, -1.93_23]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , a , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'''Saving model and image processor to {pytorch_dump_folder_path}''' ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) image_processor.save_pretrained(a ) if push_to_hub: print('''Pushing model and image processor to the hub...''' ) model.push_to_hub(F'''nielsr/{model_name}''' ) image_processor.push_to_hub(F'''nielsr/{model_name}''' ) if __name__ == "__main__": _lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="maskformer-swin-tiny-ade", type=str, help=("Name of the MaskFormer model you'd like to convert",), ) parser.add_argument( "--checkpoint_path", default="/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl", type=str, help="Path to the original state dict (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) _lowerCAmelCase = parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
98
1
import argparse import math import os from copy import deepcopy import torch from audio_diffusion.models import DiffusionAttnUnetaD from diffusion import sampling from torch import nn from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel lowercase : Any = { 'gwf-440k': { 'url': 'https://model-server.zqevans2.workers.dev/gwf-440k.ckpt', 'sample_rate': 48000, 'sample_size': 65536, }, 'jmann-small-190k': { 'url': 'https://model-server.zqevans2.workers.dev/jmann-small-190k.ckpt', 'sample_rate': 48000, 'sample_size': 65536, }, 'jmann-large-580k': { 'url': 'https://model-server.zqevans2.workers.dev/jmann-large-580k.ckpt', 'sample_rate': 48000, 'sample_size': 131072, }, 'maestro-uncond-150k': { 'url': 'https://model-server.zqevans2.workers.dev/maestro-uncond-150k.ckpt', 'sample_rate': 16000, 'sample_size': 65536, }, 'unlocked-uncond-250k': { 'url': 'https://model-server.zqevans2.workers.dev/unlocked-uncond-250k.ckpt', 'sample_rate': 16000, 'sample_size': 65536, }, 'honk-140k': { 'url': 'https://model-server.zqevans2.workers.dev/honk-140k.ckpt', 'sample_rate': 16000, 'sample_size': 65536, }, } def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : str , _lowerCamelCase : List[Any]) -> str: '''simple docstring''' return torch.atana(_lowerCamelCase , _lowerCamelCase) / math.pi * 2 def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[Any]) -> Dict: '''simple docstring''' __UpperCamelCase : Any = torch.sin(t * math.pi / 2) ** 2 __UpperCamelCase : List[Any] = (1 - sigma**2) ** 0.5 return alpha_sigma_to_t(_lowerCamelCase , _lowerCamelCase) class lowerCamelCase__ ( __lowercase): '''simple docstring''' pass class lowerCamelCase__ ( nn.Module): '''simple docstring''' def __init__( self :Union[str, Any] , a :int ) -> List[str]: super().__init__() __UpperCamelCase : List[Any] = DiffusionAttnUnetaD(a , n_attn_layers=4 ) __UpperCamelCase : str = deepcopy(self.diffusion ) __UpperCamelCase : Tuple = torch.quasirandom.SobolEngine(1 , scramble=a ) def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[Any]) -> Optional[Any]: '''simple docstring''' __UpperCamelCase : Dict = MODELS_MAP[model_name]["url"] os.system(F'wget {url} ./') return F'./{model_name}.ckpt' lowercase : Optional[int] = { '1': 'resnets.0', '2': 'attentions.0', '3': 'resnets.1', '4': 'attentions.1', '5': 'resnets.2', '6': 'attentions.2', } lowercase : List[Any] = { '8': 'resnets.0', '9': 'attentions.0', '10': 'resnets.1', '11': 'attentions.1', '12': 'resnets.2', '13': 'attentions.2', } lowercase : Union[str, Any] = { '1': 'resnets.0', '2': 'attentions.0', '3': 'resnets.1', '4': 'attentions.1', '5': 'resnets.2', '6': 'attentions.2', '8': 'resnets.3', '9': 'attentions.3', '10': 'resnets.4', '11': 'attentions.4', '12': 'resnets.5', '13': 'attentions.5', } lowercase : str = { '0': 'resnets.0', '1': 'resnets.1', '2': 'resnets.2', '4': 'resnets.0', '5': 'resnets.1', '6': 'resnets.2', } lowercase : Optional[int] = { 'skip': 'conv_skip', 'main.0': 'conv_1', 'main.1': 'group_norm_1', 'main.3': 'conv_2', 'main.4': 'group_norm_2', } lowercase : Any = { 'norm': 'group_norm', 'qkv_proj': ['query', 'key', 'value'], 'out_proj': ['proj_attn'], } def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[int]) -> Any: '''simple docstring''' if name.startswith("skip"): return name.replace("skip" , RES_CONV_MAP["skip"]) # name has to be of format main.{digit} if not name.startswith("main."): raise ValueError(F'ResConvBlock error with {name}') return name.replace(name[:6] , RES_CONV_MAP[name[:6]]) def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Any) -> Any: '''simple docstring''' for key, value in ATTN_MAP.items(): if name.startswith(_lowerCamelCase) and not isinstance(_lowerCamelCase , _lowerCamelCase): return name.replace(_lowerCamelCase , _lowerCamelCase) elif name.startswith(_lowerCamelCase): return [name.replace(_lowerCamelCase , _lowerCamelCase) for v in value] raise ValueError(F'Attn error with {name}') def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : str , _lowerCamelCase : Any=13) -> Tuple: '''simple docstring''' __UpperCamelCase : Dict = input_string if string.split(".")[0] == "timestep_embed": return string.replace("timestep_embed" , "time_proj") __UpperCamelCase : int = 0 if string.startswith("net.3."): depth += 1 __UpperCamelCase : Optional[Any] = string[6:] elif string.startswith("net."): __UpperCamelCase : Union[str, Any] = string[4:] while string.startswith("main.7."): depth += 1 __UpperCamelCase : Optional[int] = string[7:] if string.startswith("main."): __UpperCamelCase : int = string[5:] # mid block if string[:2].isdigit(): __UpperCamelCase : List[Any] = string[:2] __UpperCamelCase : Any = string[2:] else: __UpperCamelCase : Dict = string[0] __UpperCamelCase : int = string[1:] if depth == max_depth: __UpperCamelCase : Union[str, Any] = MID_NUM_TO_LAYER[layer_num] __UpperCamelCase : Optional[Any] = "mid_block" elif depth > 0 and int(_lowerCamelCase) < 7: __UpperCamelCase : str = DOWN_NUM_TO_LAYER[layer_num] __UpperCamelCase : Dict = F'down_blocks.{depth}' elif depth > 0 and int(_lowerCamelCase) > 7: __UpperCamelCase : Dict = UP_NUM_TO_LAYER[layer_num] __UpperCamelCase : Union[str, Any] = F'up_blocks.{max_depth - depth - 1}' elif depth == 0: __UpperCamelCase : List[str] = DEPTH_0_TO_LAYER[layer_num] __UpperCamelCase : str = F'up_blocks.{max_depth - 1}' if int(_lowerCamelCase) > 3 else "down_blocks.0" if not string_left.startswith("."): raise ValueError(F'Naming error with {input_string} and string_left: {string_left}.') __UpperCamelCase : Optional[int] = string_left[1:] if "resnets" in new_layer: __UpperCamelCase : List[str] = convert_resconv_naming(_lowerCamelCase) elif "attentions" in new_layer: __UpperCamelCase : Union[str, Any] = convert_attn_naming(_lowerCamelCase) __UpperCamelCase : Tuple = new_string_left if not isinstance(_lowerCamelCase , _lowerCamelCase): __UpperCamelCase : List[str] = prefix + "." + new_layer + "." + string_left else: __UpperCamelCase : str = [prefix + "." + new_layer + "." + s for s in string_left] return new_string def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : str) -> Any: '''simple docstring''' __UpperCamelCase : Union[str, Any] = {} for k, v in state_dict.items(): if k.endswith("kernel"): # up- and downsample layers, don't have trainable weights continue __UpperCamelCase : Dict = rename(_lowerCamelCase) # check if we need to transform from Conv => Linear for attention if isinstance(_lowerCamelCase , _lowerCamelCase): __UpperCamelCase : str = transform_conv_attns(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase) else: __UpperCamelCase : int = v return new_state_dict def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[int] , _lowerCamelCase : int , _lowerCamelCase : int) -> Tuple: '''simple docstring''' if len(_lowerCamelCase) == 1: if len(v.shape) == 3: # weight __UpperCamelCase : int = v[:, :, 0] else: # bias __UpperCamelCase : Union[str, Any] = v else: # qkv matrices __UpperCamelCase : str = v.shape[0] __UpperCamelCase : int = trippled_shape // 3 for i in range(3): if len(v.shape) == 3: __UpperCamelCase : List[str] = v[i * single_shape : (i + 1) * single_shape, :, 0] else: __UpperCamelCase : str = v[i * single_shape : (i + 1) * single_shape] return new_state_dict def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : Optional[int]) -> int: '''simple docstring''' __UpperCamelCase : Optional[Any] = torch.device("cuda" if torch.cuda.is_available() else "cpu") __UpperCamelCase : int = args.model_path.split("/")[-1].split(".")[0] if not os.path.isfile(args.model_path): assert ( model_name == args.model_path ), F'Make sure to provide one of the official model names {MODELS_MAP.keys()}' __UpperCamelCase : str = download(_lowerCamelCase) __UpperCamelCase : Optional[Any] = MODELS_MAP[model_name]["sample_rate"] __UpperCamelCase : Tuple = MODELS_MAP[model_name]["sample_size"] __UpperCamelCase : Union[str, Any] = Object() __UpperCamelCase : Optional[int] = sample_size __UpperCamelCase : str = sample_rate __UpperCamelCase : Optional[Any] = 0 __UpperCamelCase : List[Any] = UNetaDModel(sample_size=_lowerCamelCase , sample_rate=_lowerCamelCase) __UpperCamelCase : List[str] = diffusers_model.state_dict() __UpperCamelCase : List[Any] = DiffusionUncond(_lowerCamelCase) orig_model.load_state_dict(torch.load(args.model_path , map_location=_lowerCamelCase)["state_dict"]) __UpperCamelCase : Any = orig_model.diffusion_ema.eval() __UpperCamelCase : Dict = orig_model.state_dict() __UpperCamelCase : int = rename_orig_weights(_lowerCamelCase) __UpperCamelCase : List[Any] = set(renamed_state_dict.keys()) - set(diffusers_state_dict.keys()) __UpperCamelCase : Any = set(diffusers_state_dict.keys()) - set(renamed_state_dict.keys()) assert len(_lowerCamelCase) == 0, F'Problem with {renamed_minus_diffusers}' assert all(k.endswith("kernel") for k in list(_lowerCamelCase)), F'Problem with {diffusers_minus_renamed}' for key, value in renamed_state_dict.items(): assert ( diffusers_state_dict[key].squeeze().shape == value.squeeze().shape ), F'Shape for {key} doesn\'t match. Diffusers: {diffusers_state_dict[key].shape} vs. {value.shape}' if key == "time_proj.weight": __UpperCamelCase : Any = value.squeeze() __UpperCamelCase : Optional[int] = value diffusers_model.load_state_dict(_lowerCamelCase) __UpperCamelCase : Union[str, Any] = 100 __UpperCamelCase : List[str] = 33 __UpperCamelCase : List[str] = IPNDMScheduler(num_train_timesteps=_lowerCamelCase) __UpperCamelCase : Tuple = torch.manual_seed(_lowerCamelCase) __UpperCamelCase : List[str] = torch.randn([1, 2, config.sample_size] , generator=_lowerCamelCase).to(_lowerCamelCase) __UpperCamelCase : str = torch.linspace(1 , 0 , steps + 1 , device=_lowerCamelCase)[:-1] __UpperCamelCase : List[str] = get_crash_schedule(_lowerCamelCase) __UpperCamelCase : Union[str, Any] = DanceDiffusionPipeline(unet=_lowerCamelCase , scheduler=_lowerCamelCase) __UpperCamelCase : Union[str, Any] = torch.manual_seed(33) __UpperCamelCase : int = pipe(num_inference_steps=_lowerCamelCase , generator=_lowerCamelCase).audios __UpperCamelCase : Optional[int] = sampling.iplms_sample(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {}) __UpperCamelCase : List[str] = generated.clamp(-1 , 1) __UpperCamelCase : str = (generated - audio).abs().sum() __UpperCamelCase : List[Any] = (generated - audio).abs().max() if args.save: pipe.save_pretrained(args.checkpoint_path) print("Diff sum" , _lowerCamelCase) print("Diff max" , _lowerCamelCase) assert diff_max < 1e-3, F'Diff max: {diff_max} is too much :-/' print(F'Conversion for {model_name} successful!') if __name__ == "__main__": lowercase : Tuple = argparse.ArgumentParser() parser.add_argument('--model_path', default=None, type=str, required=True, help='Path to the model to convert.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument('--checkpoint_path', default=None, type=str, required=True, help='Path to the output model.') lowercase : str = parser.parse_args() main(args)
232
class lowerCamelCase__ : '''simple docstring''' def __init__( self :Union[str, Any] ) -> None: __UpperCamelCase : dict[str, TrieNode] = {} # Mapping from char to TrieNode __UpperCamelCase : List[str] = False def _lowerCamelCase ( self :Any , a :list[str] ) -> None: for word in words: self.insert(a ) def _lowerCamelCase ( self :List[str] , a :str ) -> None: __UpperCamelCase : Dict = self for char in word: if char not in curr.nodes: __UpperCamelCase : List[Any] = TrieNode() __UpperCamelCase : List[Any] = curr.nodes[char] __UpperCamelCase : Union[str, Any] = True def _lowerCamelCase ( self :Optional[int] , a :str ) -> bool: __UpperCamelCase : Union[str, Any] = self for char in word: if char not in curr.nodes: return False __UpperCamelCase : Union[str, Any] = curr.nodes[char] return curr.is_leaf def _lowerCamelCase ( self :Any , a :str ) -> None: def _delete(a :TrieNode , a :str , a :int ) -> bool: if index == len(a ): # If word does not exist if not curr.is_leaf: return False __UpperCamelCase : str = False return len(curr.nodes ) == 0 __UpperCamelCase : List[Any] = word[index] __UpperCamelCase : Optional[int] = curr.nodes.get(a ) # If char not in current trie node if not char_node: return False # Flag to check if node can be deleted __UpperCamelCase : int = _delete(a , a , index + 1 ) if delete_curr: del curr.nodes[char] return len(curr.nodes ) == 0 return delete_curr _delete(self , a , 0 ) def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : TrieNode , _lowerCamelCase : str) -> None: '''simple docstring''' if node.is_leaf: print(_lowerCamelCase , end=" ") for key, value in node.nodes.items(): print_words(_lowerCamelCase , word + key) def _SCREAMING_SNAKE_CASE ( ) -> bool: '''simple docstring''' __UpperCamelCase : int = "banana bananas bandana band apple all beast".split() __UpperCamelCase : Union[str, Any] = TrieNode() root.insert_many(_lowerCamelCase) # print_words(root, "") assert all(root.find(_lowerCamelCase) for word in words) assert root.find("banana") assert not root.find("bandanas") assert not root.find("apps") assert root.find("apple") assert root.find("all") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def _SCREAMING_SNAKE_CASE ( _lowerCamelCase : str , _lowerCamelCase : bool) -> None: '''simple docstring''' print(str(_lowerCamelCase) , "works!" if passes else "doesn't work :(") def _SCREAMING_SNAKE_CASE ( ) -> None: '''simple docstring''' assert test_trie() def _SCREAMING_SNAKE_CASE ( ) -> None: '''simple docstring''' print_results("Testing trie functionality" , test_trie()) if __name__ == "__main__": main()
232
1
'''simple docstring''' import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() __UpperCAmelCase =logging.get_logger(__name__) def __lowerCAmelCase ( UpperCamelCase__ ) -> Dict: __lowerCamelCase = SwinConfig.from_pretrained( '''microsoft/swin-tiny-patch4-window7-224''' , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) __lowerCamelCase = MaskFormerConfig(backbone_config=UpperCamelCase__ ) __lowerCamelCase = '''huggingface/label-files''' if "ade20k-full" in model_name: # this should be ok __lowerCamelCase = 8_47 __lowerCamelCase = '''maskformer-ade20k-full-id2label.json''' elif "ade" in model_name: # this should be ok __lowerCamelCase = 1_50 __lowerCamelCase = '''ade20k-id2label.json''' elif "coco-stuff" in model_name: # this should be ok __lowerCamelCase = 1_71 __lowerCamelCase = '''maskformer-coco-stuff-id2label.json''' elif "coco" in model_name: # TODO __lowerCamelCase = 1_33 __lowerCamelCase = '''coco-panoptic-id2label.json''' elif "cityscapes" in model_name: # this should be ok __lowerCamelCase = 19 __lowerCamelCase = '''cityscapes-id2label.json''' elif "vistas" in model_name: # this should be ok __lowerCamelCase = 65 __lowerCamelCase = '''mapillary-vistas-id2label.json''' __lowerCamelCase = json.load(open(hf_hub_download(UpperCamelCase__ , UpperCamelCase__ , repo_type='''dataset''' ) , '''r''' ) ) __lowerCamelCase = {int(UpperCamelCase__ ): v for k, v in idalabel.items()} return config def __lowerCAmelCase ( UpperCamelCase__ ) -> List[str]: __lowerCamelCase = [] # stem # fmt: off rename_keys.append(('''backbone.patch_embed.proj.weight''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.proj.bias''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''model.pixel_level_module.encoder.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''model.pixel_level_module.encoder.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm1.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm1.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.relative_position_index""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.proj.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.proj.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm2.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm2.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc1.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc1.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc2.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc2.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias""") ) if i < 3: rename_keys.append((f"""backbone.layers.{i}.downsample.reduction.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight""") ) rename_keys.append((f"""backbone.layers.{i}.downsample.norm.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight""") ) rename_keys.append((f"""backbone.layers.{i}.downsample.norm.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias""") ) rename_keys.append((f"""backbone.norm{i}.weight""", f"""model.pixel_level_module.encoder.hidden_states_norms.{i}.weight""") ) rename_keys.append((f"""backbone.norm{i}.bias""", f"""model.pixel_level_module.encoder.hidden_states_norms.{i}.bias""") ) # FPN rename_keys.append(('''sem_seg_head.layer_4.weight''', '''model.pixel_level_module.decoder.fpn.stem.0.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.weight''', '''model.pixel_level_module.decoder.fpn.stem.1.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.bias''', '''model.pixel_level_module.decoder.fpn.stem.1.bias''') ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((f"""sem_seg_head.adapter_{source_index}.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight""") ) rename_keys.append((f"""sem_seg_head.adapter_{source_index}.norm.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight""") ) rename_keys.append((f"""sem_seg_head.adapter_{source_index}.norm.bias""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.norm.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.norm.bias""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias""") ) rename_keys.append(('''sem_seg_head.mask_features.weight''', '''model.pixel_level_module.decoder.mask_projection.weight''') ) rename_keys.append(('''sem_seg_head.mask_features.bias''', '''model.pixel_level_module.decoder.mask_projection.bias''') ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight""", f"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias""", f"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias""") ) # cross-attention out projection rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias""") ) # MLP 1 rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight""", f"""model.transformer_module.decoder.layers.{idx}.fc1.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias""", f"""model.transformer_module.decoder.layers.{idx}.fc1.bias""") ) # MLP 2 rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight""", f"""model.transformer_module.decoder.layers.{idx}.fc2.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias""", f"""model.transformer_module.decoder.layers.{idx}.fc2.bias""") ) # layernorm 1 (self-attention layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight""", f"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias""", f"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias""") ) # layernorm 2 (cross-attention layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias""") ) # layernorm 3 (final layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight""", f"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias""", f"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias""") ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.weight''', '''model.transformer_module.decoder.layernorm.weight''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.bias''', '''model.transformer_module.decoder.layernorm.bias''') ) # heads on top rename_keys.append(('''sem_seg_head.predictor.query_embed.weight''', '''model.transformer_module.queries_embedder.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.weight''', '''model.transformer_module.input_projection.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.bias''', '''model.transformer_module.input_projection.bias''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.weight''', '''class_predictor.weight''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.bias''', '''class_predictor.bias''') ) for i in range(3 ): rename_keys.append((f"""sem_seg_head.predictor.mask_embed.layers.{i}.weight""", f"""mask_embedder.{i}.0.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.mask_embed.layers.{i}.bias""", f"""mask_embedder.{i}.0.bias""") ) # fmt: on return rename_keys def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: __lowerCamelCase = dct.pop(UpperCamelCase__ ) __lowerCamelCase = val def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> Optional[int]: __lowerCamelCase = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __lowerCamelCase = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f"""backbone.layers.{i}.blocks.{j}.attn.qkv.weight""" ) __lowerCamelCase = state_dict.pop(f"""backbone.layers.{i}.blocks.{j}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[:dim, :] __lowerCamelCase = in_proj_bias[: dim] __lowerCamelCase = in_proj_weight[ dim : dim * 2, : ] __lowerCamelCase = in_proj_bias[ dim : dim * 2 ] __lowerCamelCase = in_proj_weight[ -dim :, : ] __lowerCamelCase = in_proj_bias[-dim :] # fmt: on def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ ) -> List[str]: # fmt: off __lowerCamelCase = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight""" ) __lowerCamelCase = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[: hidden_size, :] __lowerCamelCase = in_proj_bias[:config.hidden_size] __lowerCamelCase = in_proj_weight[hidden_size : hidden_size * 2, :] __lowerCamelCase = in_proj_bias[hidden_size : hidden_size * 2] __lowerCamelCase = in_proj_weight[-hidden_size :, :] __lowerCamelCase = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) __lowerCamelCase = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight""" ) __lowerCamelCase = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict __lowerCamelCase = in_proj_weight[: hidden_size, :] __lowerCamelCase = in_proj_bias[:config.hidden_size] __lowerCamelCase = in_proj_weight[hidden_size : hidden_size * 2, :] __lowerCamelCase = in_proj_bias[hidden_size : hidden_size * 2] __lowerCamelCase = in_proj_weight[-hidden_size :, :] __lowerCamelCase = in_proj_bias[-hidden_size :] # fmt: on def __lowerCAmelCase ( ) -> torch.Tensor: __lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' __lowerCamelCase = Image.open(requests.get(UpperCamelCase__ , stream=UpperCamelCase__ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = False ) -> str: __lowerCamelCase = get_maskformer_config(UpperCamelCase__ ) # load original state_dict with open(UpperCamelCase__ , '''rb''' ) as f: __lowerCamelCase = pickle.load(UpperCamelCase__ ) __lowerCamelCase = data['''model'''] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys __lowerCamelCase = create_rename_keys(UpperCamelCase__ ) for src, dest in rename_keys: rename_key(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) read_in_swin_q_k_v(UpperCamelCase__ , config.backbone_config ) read_in_decoder_q_k_v(UpperCamelCase__ , UpperCamelCase__ ) # update to torch tensors for key, value in state_dict.items(): __lowerCamelCase = torch.from_numpy(UpperCamelCase__ ) # load 🤗 model __lowerCamelCase = MaskFormerForInstanceSegmentation(UpperCamelCase__ ) model.eval() for name, param in model.named_parameters(): print(UpperCamelCase__ , param.shape ) __lowerCamelCase , __lowerCamelCase = model.load_state_dict(UpperCamelCase__ , strict=UpperCamelCase__ ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(UpperCamelCase__ ) == 0, f"""Unexpected keys: {unexpected_keys}""" # verify results __lowerCamelCase = prepare_img() if "vistas" in model_name: __lowerCamelCase = 65 elif "cityscapes" in model_name: __lowerCamelCase = 6_55_35 else: __lowerCamelCase = 2_55 __lowerCamelCase = True if '''ade''' in model_name else False __lowerCamelCase = MaskFormerImageProcessor(ignore_index=UpperCamelCase__ , reduce_labels=UpperCamelCase__ ) __lowerCamelCase = image_processor(UpperCamelCase__ , return_tensors='''pt''' ) __lowerCamelCase = model(**UpperCamelCase__ ) print('''Logits:''' , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": __lowerCamelCase = torch.tensor( [[3.6_3_5_3, -4.4_7_7_0, -2.6_0_6_5], [0.5_0_8_1, -4.2_3_9_4, -3.5_3_4_3], [2.1_9_0_9, -5.0_3_5_3, -1.9_3_2_3]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , UpperCamelCase__ , atol=1E-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f"""Saving model and image processor to {pytorch_dump_folder_path}""" ) Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) image_processor.save_pretrained(UpperCamelCase__ ) if push_to_hub: print('''Pushing model and image processor to the hub...''' ) model.push_to_hub(f"""nielsr/{model_name}""" ) image_processor.push_to_hub(f"""nielsr/{model_name}""" ) if __name__ == "__main__": __UpperCAmelCase =argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="maskformer-swin-tiny-ade", type=str, help=("Name of the MaskFormer model you'd like to convert",), ) parser.add_argument( "--checkpoint_path", default="/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl", type=str, help="Path to the original state dict (.pth file).", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __UpperCAmelCase =parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
237
'''simple docstring''' from __future__ import annotations from typing import Any def __lowerCAmelCase ( UpperCamelCase__ ) -> None: create_state_space_tree(UpperCamelCase__ , [] , 0 ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> None: if index == len(UpperCamelCase__ ): print(UpperCamelCase__ ) return create_state_space_tree(UpperCamelCase__ , UpperCamelCase__ , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(UpperCamelCase__ , UpperCamelCase__ , index + 1 ) current_subsequence.pop() if __name__ == "__main__": __UpperCAmelCase =[3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
237
1
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def a_ ( __lowercase : List[Any]=None , __lowercase : List[Any]=None ) -> Tuple: return field(default_factory=lambda: default , metadata=__lowercase ) @dataclass class SCREAMING_SNAKE_CASE__ : '''simple docstring''' _UpperCAmelCase : str = field( metadata={"help": "The csv file to plot."} ,) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={"help": "Whether to plot along batch size or sequence length. Defaults to sequence length."} ,) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={"help": "Whether the csv file has time results or memory results. Defaults to memory results."} ,) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={"help": "Disable logarithmic scale when plotting"} ,) _UpperCAmelCase : bool = field( default=UpperCAmelCase ,metadata={ "help": "Whether the csv file has training results or inference results. Defaults to inference results." } ,) _UpperCAmelCase : Optional[str] = field( default=UpperCAmelCase ,metadata={"help": "Filename under which the plot will be saved. If unused no plot is saved."} ,) _UpperCAmelCase : Optional[List[str]] = list_field( default=UpperCAmelCase ,metadata={"help": "List of model names that are used instead of the ones in the csv file."} ) def a_ ( __lowercase : Tuple ) -> Optional[int]: try: int(__lowercase ) return True except ValueError: return False def a_ ( __lowercase : Tuple ) -> List[str]: try: float(__lowercase ) return True except ValueError: return False class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : Dict , lowercase : Dict ): '''simple docstring''' _snake_case = args _snake_case = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline='' ) as csv_file: _snake_case = csv.DictReader(lowercase ) for row in reader: _snake_case = row['model'] self.result_dict[model_name]["bsz"].append(int(row['batch_size'] ) ) self.result_dict[model_name]["seq_len"].append(int(row['sequence_length'] ) ) if can_convert_to_int(row['result'] ): # value is not None _snake_case = int(row['result'] ) elif can_convert_to_float(row['result'] ): # value is not None _snake_case = float(row['result'] ) def A ( self : Dict ): '''simple docstring''' _snake_case , _snake_case = plt.subplots() _snake_case = 'Time usage' if self.args.is_time else 'Memory usage' _snake_case = title_str + ' for training' if self.args.is_train else title_str + ' for inference' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('log' ) ax.set_yscale('log' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): _snake_case = sorted(set(self.result_dict[model_name]['bsz'] ) ) _snake_case = sorted(set(self.result_dict[model_name]['seq_len'] ) ) _snake_case = self.result_dict[model_name]['result'] ((_snake_case) , (_snake_case)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) _snake_case = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: _snake_case = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=lowercase , ) else: _snake_case = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((_snake_case) , (_snake_case)) = ( ('batch_size', 'len') if self.args.plot_along_batch else ('in #tokens', 'bsz') ) _snake_case = np.asarray(lowercase , lowercase )[: len(lowercase )] plt.scatter( lowercase , lowercase , label=f'''{label_model_name} - {inner_loop_label}: {inner_loop_value}''' ) plt.plot(lowercase , lowercase , '--' ) title_str += f''' {label_model_name} vs.''' _snake_case = title_str[:-4] _snake_case = 'Time in s' if self.args.is_time else 'Memory in MB' # plot plt.title(lowercase ) plt.xlabel(lowercase ) plt.ylabel(lowercase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def a_ ( ) -> Dict: _snake_case = HfArgumentParser(__lowercase ) _snake_case = parser.parse_args_into_dataclasses()[0] _snake_case = Plot(args=__lowercase ) plot.plot() if __name__ == "__main__": main()
282
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: _lowerCamelCase : int = None _lowerCamelCase : List[str] = logging.get_logger(__name__) _lowerCamelCase : Tuple = '''▁''' _lowerCamelCase : Optional[Any] = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} _lowerCamelCase : Any = { '''vocab_file''': {'''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'''}, '''tokenizer_file''': { '''google/pegasus-xsum''': '''https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json''' }, } _lowerCamelCase : Optional[int] = { '''google/pegasus-xsum''': 512, } class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _UpperCAmelCase : int = VOCAB_FILES_NAMES _UpperCAmelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Any = PegasusTokenizer _UpperCAmelCase : Dict = ["input_ids", "attention_mask"] def __init__( self : Tuple , lowercase : str=None , lowercase : Any=None , lowercase : List[Any]="<pad>" , lowercase : List[Any]="</s>" , lowercase : Tuple="<unk>" , lowercase : Any="<mask_2>" , lowercase : List[str]="<mask_1>" , lowercase : List[Any]=None , lowercase : Dict=103 , **lowercase : Optional[Any] , ): '''simple docstring''' _snake_case = offset if additional_special_tokens is not None: if not isinstance(lowercase , lowercase ): raise TypeError( f'''additional_special_tokens should be of type {type(lowercase )}, but is''' f''' {type(lowercase )}''' ) _snake_case = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'''<unk_{i}>''' for i in range(len(lowercase ) , self.offset - 1 ) ] if len(set(lowercase ) ) != len(lowercase ): raise ValueError( 'Please make sure that the provided additional_special_tokens do not contain an incorrectly' f''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) _snake_case = additional_special_tokens_extended else: _snake_case = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'''<unk_{i}>''' for i in range(2 , self.offset )] super().__init__( lowercase , tokenizer_file=lowercase , pad_token=lowercase , eos_token=lowercase , unk_token=lowercase , mask_token=lowercase , mask_token_sent=lowercase , offset=lowercase , additional_special_tokens=lowercase , **lowercase , ) _snake_case = vocab_file _snake_case = False if not self.vocab_file else True def A ( self : List[str] , lowercase : Optional[int] ): '''simple docstring''' _snake_case = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( 'There should be 3 special tokens: mask_token, pad_token, and eos_token +' f''' {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}''' ) return [1 if x in all_special_ids else 0 for x in seq] def A ( self : List[Any] , lowercase : List , lowercase : Optional[List] = None , lowercase : bool = False ): '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(lowercase ) elif token_ids_a is None: return self._special_token_mask(lowercase ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def A ( self : Any , lowercase : Tuple , lowercase : Any=None ): '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def A ( self : int , lowercase : str , lowercase : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(lowercase ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return _snake_case = os.path.join( lowercase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase ): copyfile(self.vocab_file , lowercase ) return (out_vocab_file,)
282
1
def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: List[str] , __lowerCamelCase: Union[str, Any] ): '''simple docstring''' lowercase_ = [1] for i in range(2 , __lowerCamelCase ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of bounds" lowercase_ = [] lowercase_ = list(range(__lowerCamelCase ) ) # Find permutation while factorials: lowercase_ = factorials.pop() lowercase_ , lowercase_ = divmod(__lowerCamelCase , __lowerCamelCase ) permutation.append(elements[number] ) elements.remove(elements[number] ) permutation.append(elements[0] ) return permutation if __name__ == "__main__": import doctest doctest.testmod()
353
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__ = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_file""": """tokenizer.json"""} # See all MVP models at https://huggingface.co/models?filter=mvp SCREAMING_SNAKE_CASE__ = { """vocab_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json""", }, """added_tokens.json""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json""", }, """merges_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt""", }, """tokenizer_file""": { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json""", }, } SCREAMING_SNAKE_CASE__ = { """RUCAIBox/mvp""": 1_0_2_4, } class __lowerCamelCase ( snake_case_ ): """simple docstring""" lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = ["input_ids", "attention_mask"] lowerCAmelCase__ = MvpTokenizer def __init__( self , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase=None , UpperCAmelCase="replace" , UpperCAmelCase="<s>" , UpperCAmelCase="</s>" , UpperCAmelCase="</s>" , UpperCAmelCase="<s>" , UpperCAmelCase="<unk>" , UpperCAmelCase="<pad>" , UpperCAmelCase="<mask>" , UpperCAmelCase=False , UpperCAmelCase=True , **UpperCAmelCase , ) -> Optional[int]: '''simple docstring''' super().__init__( UpperCAmelCase , UpperCAmelCase , tokenizer_file=UpperCAmelCase , errors=UpperCAmelCase , bos_token=UpperCAmelCase , eos_token=UpperCAmelCase , sep_token=UpperCAmelCase , cls_token=UpperCAmelCase , unk_token=UpperCAmelCase , pad_token=UpperCAmelCase , mask_token=UpperCAmelCase , add_prefix_space=UpperCAmelCase , trim_offsets=UpperCAmelCase , **UpperCAmelCase , ) lowercase_ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("add_prefix_space" , UpperCAmelCase ) != add_prefix_space: lowercase_ = getattr(UpperCAmelCase , pre_tok_state.pop("type" ) ) lowercase_ = add_prefix_space lowercase_ = pre_tok_class(**UpperCAmelCase ) lowercase_ = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` lowercase_ = "post_processor" lowercase_ = getattr(self.backend_tokenizer , UpperCAmelCase , UpperCAmelCase ) if tokenizer_component_instance: lowercase_ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase_ = tuple(state["sep"] ) if "cls" in state: lowercase_ = tuple(state["cls"] ) lowercase_ = False if state.get("add_prefix_space" , UpperCAmelCase ) != add_prefix_space: lowercase_ = add_prefix_space lowercase_ = True if state.get("trim_offsets" , UpperCAmelCase ) != trim_offsets: lowercase_ = trim_offsets lowercase_ = True if changes_to_apply: lowercase_ = getattr(UpperCAmelCase , state.pop("type" ) ) lowercase_ = component_class(**UpperCAmelCase ) setattr(self.backend_tokenizer , UpperCAmelCase , UpperCAmelCase ) @property def A__ ( self ) -> str: '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("Using mask_token, but it is not set yet." ) return None return str(self._mask_token ) @mask_token.setter def A__ ( self , UpperCAmelCase ) -> int: '''simple docstring''' lowercase_ = AddedToken(UpperCAmelCase , lstrip=UpperCAmelCase , rstrip=UpperCAmelCase ) if isinstance(UpperCAmelCase , UpperCAmelCase ) else value lowercase_ = value def A__ ( self , *UpperCAmelCase , **UpperCAmelCase ) -> BatchEncoding: '''simple docstring''' lowercase_ = kwargs.get("is_split_into_words" , UpperCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCAmelCase , **UpperCAmelCase ) def A__ ( self , *UpperCAmelCase , **UpperCAmelCase ) -> BatchEncoding: '''simple docstring''' lowercase_ = kwargs.get("is_split_into_words" , UpperCAmelCase ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCAmelCase , **UpperCAmelCase ) def A__ ( self , UpperCAmelCase , UpperCAmelCase = None ) -> Tuple[str]: '''simple docstring''' lowercase_ = self._tokenizer.model.save(UpperCAmelCase , name=UpperCAmelCase ) return tuple(UpperCAmelCase ) def A__ ( self , UpperCAmelCase , UpperCAmelCase=None ) -> Tuple: '''simple docstring''' lowercase_ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def A__ ( self , UpperCAmelCase , UpperCAmelCase = None ) -> List[int]: '''simple docstring''' lowercase_ = [self.sep_token_id] lowercase_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
297
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCAmelCase : Any = logging.get_logger(__name__) _UpperCAmelCase : Dict = { """vinvino02/glpn-kitti""": """https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json""", # See all GLPN models at https://huggingface.co/models?filter=glpn } class lowerCAmelCase ( __UpperCamelCase ): UpperCAmelCase__ = """glpn""" def __init__( self : int , UpperCAmelCase : List[str]=3 , UpperCAmelCase : Tuple=4 , UpperCAmelCase : List[Any]=[2, 2, 2, 2] , UpperCAmelCase : int=[8, 4, 2, 1] , UpperCAmelCase : Tuple=[32, 64, 160, 256] , UpperCAmelCase : str=[7, 3, 3, 3] , UpperCAmelCase : Union[str, Any]=[4, 2, 2, 2] , UpperCAmelCase : List[Any]=[1, 2, 5, 8] , UpperCAmelCase : Tuple=[4, 4, 4, 4] , UpperCAmelCase : Optional[int]="gelu" , UpperCAmelCase : Any=0.0 , UpperCAmelCase : Any=0.0 , UpperCAmelCase : int=0.0_2 , UpperCAmelCase : str=0.1 , UpperCAmelCase : List[str]=1e-6 , UpperCAmelCase : Dict=64 , UpperCAmelCase : Optional[int]=10 , UpperCAmelCase : Any=-1 , **UpperCAmelCase : List[Any] , ) -> str: super().__init__(**UpperCAmelCase ) lowerCamelCase__ : int = num_channels lowerCamelCase__ : str = num_encoder_blocks lowerCamelCase__ : Any = depths lowerCamelCase__ : int = sr_ratios lowerCamelCase__ : List[Any] = hidden_sizes lowerCamelCase__ : List[str] = patch_sizes lowerCamelCase__ : int = strides lowerCamelCase__ : List[str] = mlp_ratios lowerCamelCase__ : Union[str, Any] = num_attention_heads lowerCamelCase__ : List[Any] = hidden_act lowerCamelCase__ : Dict = hidden_dropout_prob lowerCamelCase__ : Dict = attention_probs_dropout_prob lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Optional[Any] = drop_path_rate lowerCamelCase__ : int = layer_norm_eps lowerCamelCase__ : List[Any] = decoder_hidden_size lowerCamelCase__ : List[Any] = max_depth lowerCamelCase__ : Tuple = head_in_index
50
import os import sys import tempfile import torch from .state import AcceleratorState from .utils import PrecisionType, PrepareForLaunch, is_mps_available, patch_environment def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=None , _UpperCAmelCase="no" , _UpperCAmelCase="29500" ) -> Tuple: lowerCamelCase__ : Dict = False lowerCamelCase__ : Dict = False if any(key.startswith('KAGGLE' ) for key in os.environ.keys() ): lowerCamelCase__ : Optional[Any] = True elif "IPython" in sys.modules: lowerCamelCase__ : Optional[Any] = 'google.colab' in str(sys.modules['IPython'].get_ipython() ) try: lowerCamelCase__ : List[str] = PrecisionType(mixed_precision.lower() ) except ValueError: raise ValueError( F"""Unknown mixed_precision mode: {args.mixed_precision.lower()}. Choose between {PrecisionType.list()}.""" ) if (in_colab or in_kaggle) and (os.environ.get('TPU_NAME' , _UpperCAmelCase ) is not None): # TPU launch import torch_xla.distributed.xla_multiprocessing as xmp if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To train on TPU in Colab or Kaggle Kernel, the `Accelerator` should only be initialized inside ' 'your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if num_processes is None: lowerCamelCase__ : Optional[Any] = 8 lowerCamelCase__ : List[str] = PrepareForLaunch(_UpperCAmelCase , distributed_type='TPU' ) print(F"""Launching a training on {num_processes} TPU cores.""" ) xmp.spawn(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) elif in_colab: # No need for a distributed launch otherwise as it's either CPU or one GPU. if torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on one CPU.' ) function(*_UpperCAmelCase ) else: if num_processes is None: raise ValueError( 'You have to specify the number of GPUs you would like to use, add `num_processes=...` to your call.' ) if num_processes > 1: # Multi-GPU launch from torch.multiprocessing import start_processes from torch.multiprocessing.spawn import ProcessRaisedException if len(AcceleratorState._shared_state ) > 0: raise ValueError( 'To launch a multi-GPU training from your notebook, the `Accelerator` should only be initialized ' 'inside your training function. Restart your notebook and make sure no cells initializes an ' '`Accelerator`.' ) if torch.cuda.is_initialized(): raise ValueError( 'To launch a multi-GPU training from your notebook, you need to avoid running any instruction ' 'using `torch.cuda` in any cell. Restart your notebook and make sure no cells use any CUDA ' 'function.' ) # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port=_UpperCAmelCase , mixed_precision=_UpperCAmelCase ): lowerCamelCase__ : Optional[int] = PrepareForLaunch(_UpperCAmelCase , distributed_type='MULTI_GPU' ) print(F"""Launching training on {num_processes} GPUs.""" ) try: start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' ) except ProcessRaisedException as e: if "Cannot re-initialize CUDA in forked subprocess" in e.args[0]: raise RuntimeError( 'CUDA has been initialized before the `notebook_launcher` could create a forked subprocess. ' 'This likely stems from an outside import causing issues once the `notebook_launcher()` is called. ' 'Please review your imports and test them when running the `notebook_launcher()` to identify ' 'which one is problematic.' ) from e else: # No need for a distributed launch otherwise as it's either CPU, GPU or MPS. if is_mps_available(): lowerCamelCase__ : int = '1' print('Launching training on MPS.' ) elif torch.cuda.is_available(): print('Launching training on one GPU.' ) else: print('Launching training on CPU.' ) function(*_UpperCAmelCase ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase=() , _UpperCAmelCase=2 ) -> Optional[Any]: from torch.multiprocessing import start_processes with tempfile.NamedTemporaryFile() as tmp_file: # torch.distributed will expect a few environment variable to be here. We set the ones common to each # process here (the other ones will be set be the launcher). with patch_environment( world_size=_UpperCAmelCase , master_addr='127.0.01' , master_port='29500' , accelerate_mixed_precision='no' , accelerate_debug_rdv_file=tmp_file.name , accelerate_use_cpu='yes' , ): lowerCamelCase__ : Optional[Any] = PrepareForLaunch(_UpperCAmelCase , debug=_UpperCAmelCase ) start_processes(_UpperCAmelCase , args=_UpperCAmelCase , nprocs=_UpperCAmelCase , start_method='fork' )
50
1
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor _lowerCamelCase : List[str] = logging.get_logger(__name__) class __snake_case (_a ): def __init__( self : Optional[Any] , *_UpperCAmelCase : str , **_UpperCAmelCase : Any ) -> None: '''simple docstring''' warnings.warn( """The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use GLPNImageProcessor instead.""" , _UpperCAmelCase , ) super().__init__(*_UpperCAmelCase , **_UpperCAmelCase )
159
def _UpperCAmelCase (UpperCamelCase_ : str , UpperCamelCase_ : str ): '''simple docstring''' _lowerCAmelCase : str = len(UpperCamelCase_ ) + 1 _lowerCAmelCase : List[Any] = len(UpperCamelCase_ ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. _lowerCAmelCase : List[Any] = [[0 for i in range(UpperCamelCase_ )] for j in range(UpperCamelCase_ )] # since string of zero length match pattern of zero length _lowerCAmelCase : Optional[int] = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , UpperCamelCase_ ): _lowerCAmelCase : Optional[Any] = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , UpperCamelCase_ ): _lowerCAmelCase : Tuple = dp[0][j - 2] if pattern[j - 1] == """*""" else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , UpperCamelCase_ ): for j in range(1 , UpperCamelCase_ ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": _lowerCAmelCase : Dict = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: _lowerCAmelCase : List[str] = 1 elif pattern[j - 2] in (input_string[i - 1], "."): _lowerCAmelCase : int = dp[i - 1][j] else: _lowerCAmelCase : List[str] = 0 else: _lowerCAmelCase : List[Any] = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") _lowerCamelCase : Any = "aab" _lowerCamelCase : List[str] = "c*a*b" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(F'''{input_string} matches the given pattern {pattern}''') else: print(F'''{input_string} does not match with the given pattern {pattern}''')
159
1
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": UpperCAmelCase = input("""Enter image url: """).strip() print(F'''Downloading image from {url} ...''') UpperCAmelCase = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image UpperCAmelCase = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] UpperCAmelCase = requests.get(image_url).content UpperCAmelCase = F'''{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg''' with open(file_name, """wb""") as fp: fp.write(image_data) print(F'''Done. Image saved to disk as {file_name}.''')
256
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments UpperCAmelCase = logging.getLogger(__name__) @dataclass class UpperCAmelCase_ ( _lowercase): snake_case__ = field( default=0.0 , metadata={'''help''': '''The label smoothing epsilon to apply (if not zero).'''}) snake_case__ = field(default=_lowercase , metadata={'''help''': '''Whether to SortishSamler or not.'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Whether to use generate to calculate generative metrics (ROUGE, BLEU).'''}) snake_case__ = field(default=_lowercase , metadata={'''help''': '''whether to use adafactor'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Encoder layer dropout probability. Goes into model.config.'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Decoder layer dropout probability. Goes into model.config.'''}) snake_case__ = field(default=_lowercase , metadata={'''help''': '''Dropout probability. Goes into model.config.'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Attention dropout probability. Goes into model.config.'''}) snake_case__ = field( default='''linear''' , metadata={'''help''': F'''Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys())}'''} , )
256
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase = logging.get_logger(__name__) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_=False ) -> Any: _a : Union[str, Any] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _a : Tuple = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=False ) -> str: for i in range(config.num_hidden_layers ): if base_model: _a : Optional[Any] = '' else: _a : Optional[int] = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _a : Union[str, Any] = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) _a : Tuple = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _a : Any = in_proj_weight[ : config.hidden_size, : ] _a : Optional[Any] = in_proj_bias[: config.hidden_size] _a : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _a : str = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _a : int = in_proj_weight[ -config.hidden_size :, : ] _a : List[str] = in_proj_bias[-config.hidden_size :] def __lowerCamelCase ( lowerCAmelCase_ ) -> Dict: _a : int = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(lowerCAmelCase_ , lowerCAmelCase_ ) def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) -> Optional[int]: _a : Dict = dct.pop(lowerCAmelCase_ ) _a : Union[str, Any] = val def __lowerCamelCase ( ) -> Optional[int]: _a : List[str] = 'http://images.cocodataset.org/val2017/000000039769.jpg' _a : Union[str, Any] = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True ) -> Dict: _a : int = ViTConfig() # patch_size if model_name[-1] == "8": _a : Tuple = 8 # set labels if required if not base_model: _a : List[Any] = 1000 _a : List[str] = 'huggingface/label-files' _a : Optional[int] = 'imagenet-1k-id2label.json' _a : Union[str, Any] = json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type='dataset' ) , 'r' ) ) _a : List[Any] = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} _a : Union[str, Any] = idalabel _a : Optional[Any] = {v: k for k, v in idalabel.items()} # size of the architecture if model_name in ["dino_vits8", "dino_vits16"]: _a : List[str] = 384 _a : str = 1536 _a : Optional[Any] = 12 _a : Optional[Any] = 6 # load original model from torch hub _a : Optional[Any] = torch.hub.load('facebookresearch/dino:main' , lowerCAmelCase_ ) original_model.eval() # load state_dict of original model, remove and rename some keys _a : Optional[int] = original_model.state_dict() if base_model: remove_classification_head_(lowerCAmelCase_ ) _a : Optional[Any] = create_rename_keys(lowerCAmelCase_ , base_model=lowerCAmelCase_ ) for src, dest in rename_keys: rename_key(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) read_in_q_k_v(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # load HuggingFace model if base_model: _a : List[Any] = ViTModel(lowerCAmelCase_ , add_pooling_layer=lowerCAmelCase_ ).eval() else: _a : Union[str, Any] = ViTForImageClassification(lowerCAmelCase_ ).eval() model.load_state_dict(lowerCAmelCase_ ) # Check outputs on an image, prepared by ViTImageProcessor _a : Tuple = ViTImageProcessor() _a : Any = image_processor(images=prepare_img() , return_tensors='pt' ) _a : int = encoding['pixel_values'] _a : Optional[Any] = model(lowerCAmelCase_ ) if base_model: _a : Dict = original_model(lowerCAmelCase_ ) assert torch.allclose(lowerCAmelCase_ , outputs.last_hidden_state[:, 0, :] , atol=1E-1 ) else: _a : Optional[Any] = original_model(lowerCAmelCase_ ) assert logits.shape == outputs.logits.shape assert torch.allclose(lowerCAmelCase_ , outputs.logits , atol=1E-3 ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(lowerCAmelCase_ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(lowerCAmelCase_ ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''dino_vitb16''', type=str, help='''Name of the model trained with DINO you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--base_model''', action='''store_true''', help='''Whether to only convert the base model (no projection head weights).''', ) parser.set_defaults(base_model=True) __lowerCAmelCase = parser.parse_args() convert_vit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.base_model)
107
'''simple docstring''' import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller __lowerCAmelCase = 3 def __lowerCamelCase ( lowerCAmelCase_ ) -> int: print('Generating primitive root of p' ) while True: _a : List[Any] = random.randrange(3 , lowerCAmelCase_ ) if pow(lowerCAmelCase_ , 2 , lowerCAmelCase_ ) == 1: continue if pow(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) == 1: continue return g def __lowerCamelCase ( lowerCAmelCase_ ) -> tuple[tuple[int, int, int, int], tuple[int, int]]: print('Generating prime p...' ) _a : int = rabin_miller.generate_large_prime(lowerCAmelCase_ ) # select large prime number. _a : List[str] = primitive_root(lowerCAmelCase_ ) # one primitive root on modulo p. _a : Any = random.randrange(3 , lowerCAmelCase_ ) # private_key -> have to be greater than 2 for safety. _a : List[Any] = cryptomath.find_mod_inverse(pow(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) , lowerCAmelCase_ ) _a : Tuple = (key_size, e_a, e_a, p) _a : str = (key_size, d) return public_key, private_key def __lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ) -> None: if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print('\nWARNING:' ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" 'Use a different name or delete these files and re-run this program.' ) sys.exit() _a , _a : Dict = generate_key(lowerCAmelCase_ ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , 'w' ) as fo: fo.write(f"""{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , 'w' ) as fo: fo.write(f"""{private_key[0]},{private_key[1]}""" ) def __lowerCamelCase ( ) -> None: print('Making key files...' ) make_key_files('elgamal' , 2048 ) print('Key files generation successful' ) if __name__ == "__main__": main()
107
1
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class lowercase__ : def __init__( self : str ,lowerCamelCase__ : str ,lowerCamelCase__ : Dict=13 ,lowerCamelCase__ : List[str]=7 ,lowerCamelCase__ : Dict=True ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : Optional[int]=True ,lowerCamelCase__ : Tuple=True ,lowerCamelCase__ : Tuple=99 ,lowerCamelCase__ : Tuple=32 ,lowerCamelCase__ : str=2 ,lowerCamelCase__ : Any=4 ,lowerCamelCase__ : List[str]=37 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : Optional[int]=0.1 ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : Dict=512 ,lowerCamelCase__ : int=16 ,lowerCamelCase__ : Any=2 ,lowerCamelCase__ : Any=0.0_2 ,lowerCamelCase__ : Any=3 ,lowerCamelCase__ : Optional[Any]=4 ,lowerCamelCase__ : List[str]=None ,lowerCamelCase__ : List[str]=1000 ,): '''simple docstring''' _UpperCamelCase : List[str] = parent _UpperCamelCase : Dict = batch_size _UpperCamelCase : Tuple = seq_length _UpperCamelCase : Union[str, Any] = is_training _UpperCamelCase : Optional[Any] = use_input_mask _UpperCamelCase : Tuple = use_token_type_ids _UpperCamelCase : Optional[int] = use_labels _UpperCamelCase : Union[str, Any] = vocab_size _UpperCamelCase : Optional[Any] = hidden_size _UpperCamelCase : Optional[Any] = num_hidden_layers _UpperCamelCase : List[str] = num_attention_heads _UpperCamelCase : Optional[Any] = intermediate_size _UpperCamelCase : Optional[Any] = hidden_act _UpperCamelCase : List[str] = hidden_dropout_prob _UpperCamelCase : int = attention_probs_dropout_prob _UpperCamelCase : List[Any] = max_position_embeddings _UpperCamelCase : List[Any] = type_vocab_size _UpperCamelCase : int = type_sequence_label_size _UpperCamelCase : Union[str, Any] = initializer_range _UpperCamelCase : str = num_labels _UpperCamelCase : Optional[Any] = num_choices _UpperCamelCase : str = scope _UpperCamelCase : Any = range_bbox def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) # convert bbox to numpy since TF does not support item assignment _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length, 4] ,self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: _UpperCamelCase : Union[str, Any] = bbox[i, j, 3] _UpperCamelCase : Optional[Any] = bbox[i, j, 1] _UpperCamelCase : Union[str, Any] = t if bbox[i, j, 2] < bbox[i, j, 0]: _UpperCamelCase : Optional[int] = bbox[i, j, 2] _UpperCamelCase : Optional[Any] = bbox[i, j, 0] _UpperCamelCase : Optional[int] = t _UpperCamelCase : Optional[int] = tf.convert_to_tensor(lowerCamelCase__ ) _UpperCamelCase : str = None if self.use_input_mask: _UpperCamelCase : Any = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : str = None if self.use_token_type_ids: _UpperCamelCase : Dict = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) _UpperCamelCase : Dict = None _UpperCamelCase : Tuple = None _UpperCamelCase : List[Any] = None if self.use_labels: _UpperCamelCase : Union[str, Any] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) _UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) _UpperCamelCase : List[str] = ids_tensor([self.batch_size] ,self.num_choices ) _UpperCamelCase : Optional[Any] = LayoutLMConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase_ ( self : int ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : int ,lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : Optional[int] = TFLayoutLMModel(config=lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = model(lowerCamelCase__ ,lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ) _UpperCamelCase : str = model(lowerCamelCase__ ,lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ) _UpperCamelCase : List[str] = model(lowerCamelCase__ ,lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCamelCase_ ( self : List[str] ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[Any] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : int ,lowerCamelCase__ : Any ,lowerCamelCase__ : Any ,lowerCamelCase__ : str ): '''simple docstring''' _UpperCamelCase : List[str] = TFLayoutLMForMaskedLM(config=lowerCamelCase__ ) _UpperCamelCase : Optional[int] = model(lowerCamelCase__ ,lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ,labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Any ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : int ,lowerCamelCase__ : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' _UpperCamelCase : int = self.num_labels _UpperCamelCase : List[str] = TFLayoutLMForSequenceClassification(config=lowerCamelCase__ ) _UpperCamelCase : Dict = model(lowerCamelCase__ ,lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCamelCase_ ( self : Any ,lowerCamelCase__ : str ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : str ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Dict ,lowerCamelCase__ : Union[str, Any] ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : int ): '''simple docstring''' _UpperCamelCase : int = self.num_labels _UpperCamelCase : Tuple = TFLayoutLMForTokenClassification(config=lowerCamelCase__ ) _UpperCamelCase : List[str] = model(lowerCamelCase__ ,lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ,labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase_ ( self : Optional[Any] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : Optional[Any] ,lowerCamelCase__ : Tuple ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : Any ,lowerCamelCase__ : Any ,lowerCamelCase__ : Optional[int] ,lowerCamelCase__ : List[str] ): '''simple docstring''' _UpperCamelCase : List[Any] = TFLayoutLMForQuestionAnswering(config=lowerCamelCase__ ) _UpperCamelCase : Optional[int] = model(lowerCamelCase__ ,lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : int = self.prepare_config_and_inputs() ( ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ( _UpperCamelCase ) , ) : int = config_and_inputs _UpperCamelCase : List[Any] = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_tf class lowercase__ ( a__ , a__ , unittest.TestCase ): lowercase__ = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) lowercase__ = ( { """feature-extraction""": TFLayoutLMModel, """fill-mask""": TFLayoutLMForMaskedLM, """text-classification""": TFLayoutLMForSequenceClassification, """token-classification""": TFLayoutLMForTokenClassification, """zero-shot""": TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) lowercase__ = False lowercase__ = True lowercase__ = 10 def UpperCamelCase_ ( self : int ): '''simple docstring''' _UpperCamelCase : List[Any] = TFLayoutLMModelTester(self ) _UpperCamelCase : str = ConfigTester(self ,config_class=lowerCamelCase__ ,hidden_size=37 ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' self.config_tester.run_common_tests() def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase__ ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase__ ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase__ ) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCamelCase : str = TFLayoutLMModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @unittest.skip('Onnx compliancy broke with TF 2.10' ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' pass def A__ ( ): _UpperCamelCase : str = tf.convert_to_tensor([[1_0_1,1_0_1_9,1_0_1_4,1_0_1_6,1_0_3_7,1_2_8_4_9,4_7_4_7,1_0_0_4,1_4_2_4_6,2_2_7_8,5_4_3_9,4_5_2_4,5_0_0_2,2_9_3_0,2_1_9_3,2_9_3_0,4_3_4_1,3_2_0_8,1_0_0_5,1_0_5_5,2_1_7_1,2_8_4_8,1_1_3_0_0,3_5_3_1,1_0_2],[1_0_1,4_0_7_0,4_0_3_4,7_0_2_0,1_0_2_4,3_0_5_8,1_0_1_5,1_0_1_3,2_8_6_1,1_0_1_3,6_0_7_0,1_9_2_7_4,2_7_7_2,6_2_0_5,2_7_8_1_4,1_6_1_4_7,1_6_1_4_7,4_3_4_3,2_0_4_7,1_0_2_8_3,1_0_9_6_9,1_4_3_8_9,1_0_1_2,2_3_3_8,1_0_2]] ) # noqa: E231 _UpperCamelCase : int = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 _UpperCamelCase : Tuple = tf.convert_to_tensor([[[0,0,0,0],[4_2_3,2_3_7,4_4_0,2_5_1],[4_2_7,2_7_2,4_4_1,2_8_7],[4_1_9,1_1_5,4_3_7,1_2_9],[9_6_1,8_8_5,9_9_2,9_1_2],[2_5_6,3_8,3_3_0,5_8],[2_5_6,3_8,3_3_0,5_8],[3_3_6,4_2,3_5_3,5_7],[3_6_0,3_9,4_0_1,5_6],[3_6_0,3_9,4_0_1,5_6],[4_1_1,3_9,4_7_1,5_9],[4_7_9,4_1,5_2_8,5_9],[5_3_3,3_9,6_3_0,6_0],[6_7,1_1_3,1_3_4,1_3_1],[1_4_1,1_1_5,2_0_9,1_3_2],[6_8,1_4_9,1_3_3,1_6_6],[1_4_1,1_4_9,1_8_7,1_6_4],[1_9_5,1_4_8,2_8_7,1_6_5],[1_9_5,1_4_8,2_8_7,1_6_5],[1_9_5,1_4_8,2_8_7,1_6_5],[2_9_5,1_4_8,3_4_9,1_6_5],[4_4_1,1_4_9,4_9_2,1_6_6],[4_9_7,1_4_9,5_4_6,1_6_4],[6_4,2_0_1,1_2_5,2_1_8],[1_0_0_0,1_0_0_0,1_0_0_0,1_0_0_0]],[[0,0,0,0],[6_6_2,1_5_0,7_5_4,1_6_6],[6_6_5,1_9_9,7_4_2,2_1_1],[5_1_9,2_1_3,5_5_4,2_2_8],[5_1_9,2_1_3,5_5_4,2_2_8],[1_3_4,4_3_3,1_8_7,4_5_4],[1_3_0,4_6_7,2_0_4,4_8_0],[1_3_0,4_6_7,2_0_4,4_8_0],[1_3_0,4_6_7,2_0_4,4_8_0],[1_3_0,4_6_7,2_0_4,4_8_0],[1_3_0,4_6_7,2_0_4,4_8_0],[3_1_4,4_6_9,3_7_6,4_8_2],[5_0_4,6_8_4,5_8_2,7_0_6],[9_4_1,8_2_5,9_7_3,9_0_0],[9_4_1,8_2_5,9_7_3,9_0_0],[9_4_1,8_2_5,9_7_3,9_0_0],[9_4_1,8_2_5,9_7_3,9_0_0],[6_1_0,7_4_9,6_5_2,7_6_5],[1_3_0,6_5_9,1_6_8,6_7_2],[1_7_6,6_5_7,2_3_7,6_7_2],[2_3_8,6_5_7,3_1_2,6_7_2],[4_4_3,6_5_3,6_2_8,6_7_2],[4_4_3,6_5_3,6_2_8,6_7_2],[7_1_6,3_0_1,8_2_5,3_1_7],[1_0_0_0,1_0_0_0,1_0_0_0,1_0_0_0]]] ) # noqa: E231 _UpperCamelCase : Tuple = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) _UpperCamelCase : str = tf.convert_to_tensor([[-1_0_0,1_0,1_0,1_0,9,1,-1_0_0,7,7,-1_0_0,7,7,4,2,5,2,8,8,-1_0_0,-1_0_0,5,0,3,2,-1_0_0],[-1_0_0,1_2,1_2,1_2,-1_0_0,1_2,1_0,-1_0_0,-1_0_0,-1_0_0,-1_0_0,1_0,1_2,9,-1_0_0,-1_0_0,-1_0_0,1_0,1_0,1_0,9,1_2,-1_0_0,1_0,-1_0_0]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class lowercase__ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Any = TFLayoutLMModel.from_pretrained('microsoft/layoutlm-base-uncased' ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Any = prepare_layoutlm_batch_inputs() # forward pass _UpperCamelCase : List[Any] = model(input_ids=lowerCamelCase__ ,bbox=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ) # test the sequence output on [0, :3, :3] _UpperCamelCase : Any = tf.convert_to_tensor( [[0.1_7_8_5, -0.1_9_4_7, -0.0_4_2_5], [-0.3_2_5_4, -0.2_8_0_7, 0.2_5_5_3], [-0.5_3_9_1, -0.3_3_2_2, 0.3_3_6_4]] ,) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] ,lowerCamelCase__ ,atol=1E-3 ) ) # test the pooled output on [1, :3] _UpperCamelCase : str = tf.convert_to_tensor([-0.6_5_8_0, -0.0_2_1_4, 0.8_5_5_2] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] ,lowerCamelCase__ ,atol=1E-3 ) ) @slow def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' # initialize model with randomly initialized sequence classification head _UpperCamelCase : str = TFLayoutLMForSequenceClassification.from_pretrained('microsoft/layoutlm-base-uncased' ,num_labels=2 ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Dict = prepare_layoutlm_batch_inputs() # forward pass _UpperCamelCase : Optional[int] = model( input_ids=lowerCamelCase__ ,bbox=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ,labels=tf.convert_to_tensor([1, 1] ) ,) # test whether we get a loss as a scalar _UpperCamelCase : List[str] = outputs.loss _UpperCamelCase : Union[str, Any] = (2,) self.assertEqual(loss.shape ,lowerCamelCase__ ) # test the shape of the logits _UpperCamelCase : Optional[Any] = outputs.logits _UpperCamelCase : List[Any] = (2, 2) self.assertEqual(logits.shape ,lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' # initialize model with randomly initialized token classification head _UpperCamelCase : int = TFLayoutLMForTokenClassification.from_pretrained('microsoft/layoutlm-base-uncased' ,num_labels=13 ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : List[str] = prepare_layoutlm_batch_inputs() # forward pass _UpperCamelCase : int = model( input_ids=lowerCamelCase__ ,bbox=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ,labels=lowerCamelCase__ ) # test the shape of the logits _UpperCamelCase : List[str] = outputs.logits _UpperCamelCase : Union[str, Any] = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape ,lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' # initialize model with randomly initialized token classification head _UpperCamelCase : Any = TFLayoutLMForQuestionAnswering.from_pretrained('microsoft/layoutlm-base-uncased' ) _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : Tuple = prepare_layoutlm_batch_inputs() # forward pass _UpperCamelCase : Union[str, Any] = model(input_ids=lowerCamelCase__ ,bbox=lowerCamelCase__ ,attention_mask=lowerCamelCase__ ,token_type_ids=lowerCamelCase__ ) # test the shape of the logits _UpperCamelCase : Optional[int] = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape ,lowerCamelCase__ ) self.assertEqual(outputs.end_logits.shape ,lowerCamelCase__ )
83
'''simple docstring''' import argparse import collections import json import os import re import string import sys import numpy as np a__ : Optional[int] = re.compile(R'\b(a|an|the)\b', re.UNICODE) a__ : int = None def _UpperCamelCase ( ) -> Dict: '''simple docstring''' UpperCamelCase__ = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=__A , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=__A , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def _UpperCamelCase ( __A ) -> Dict: '''simple docstring''' UpperCamelCase__ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: UpperCamelCase__ = bool(qa["answers"]["text"] ) return qid_to_has_ans def _UpperCamelCase ( __A ) -> Optional[Any]: '''simple docstring''' def remove_articles(__A ): return ARTICLES_REGEX.sub(" " , __A ) def white_space_fix(__A ): return " ".join(text.split() ) def remove_punc(__A ): UpperCamelCase__ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__A ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__A ) ) ) ) def _UpperCamelCase ( __A ) -> int: '''simple docstring''' if not s: return [] return normalize_answer(__A ).split() def _UpperCamelCase ( __A , __A ) -> List[Any]: '''simple docstring''' return int(normalize_answer(__A ) == normalize_answer(__A ) ) def _UpperCamelCase ( __A , __A ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ = get_tokens(__A ) UpperCamelCase__ = get_tokens(__A ) UpperCamelCase__ = collections.Counter(__A ) & collections.Counter(__A ) UpperCamelCase__ = sum(common.values() ) if len(__A ) == 0 or len(__A ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 UpperCamelCase__ = 1.0 * num_same / len(__A ) UpperCamelCase__ = 1.0 * num_same / len(__A ) UpperCamelCase__ = (2 * precision * recall) / (precision + recall) return fa def _UpperCamelCase ( __A , __A ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ = {} UpperCamelCase__ = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: UpperCamelCase__ = qa["id"] UpperCamelCase__ = [t for t in qa["answers"]["text"] if normalize_answer(__A )] if not gold_answers: # For unanswerable questions, only correct answer is empty string UpperCamelCase__ = [""] if qid not in preds: print(F'''Missing prediction for {qid}''' ) continue UpperCamelCase__ = preds[qid] # Take max over all gold answers UpperCamelCase__ = max(compute_exact(__A , __A ) for a in gold_answers ) UpperCamelCase__ = max(compute_fa(__A , __A ) for a in gold_answers ) return exact_scores, fa_scores def _UpperCamelCase ( __A , __A , __A , __A ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ = {} for qid, s in scores.items(): UpperCamelCase__ = na_probs[qid] > na_prob_thresh if pred_na: UpperCamelCase__ = float(not qid_to_has_ans[qid] ) else: UpperCamelCase__ = s return new_scores def _UpperCamelCase ( __A , __A , __A=None ) -> List[Any]: '''simple docstring''' if not qid_list: UpperCamelCase__ = len(__A ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: UpperCamelCase__ = len(__A ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def _UpperCamelCase ( __A , __A , __A ) -> Optional[int]: '''simple docstring''' for k in new_eval: UpperCamelCase__ = new_eval[k] def _UpperCamelCase ( __A , __A , __A , __A ) -> Optional[int]: '''simple docstring''' plt.step(__A , __A , color="b" , alpha=0.2 , where="post" ) plt.fill_between(__A , __A , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(__A ) plt.savefig(__A ) plt.clf() def _UpperCamelCase ( __A , __A , __A , __A , __A=None , __A=None ) -> Any: '''simple docstring''' UpperCamelCase__ = sorted(__A , key=lambda __A : na_probs[k] ) UpperCamelCase__ = 0.0 UpperCamelCase__ = 1.0 UpperCamelCase__ = 0.0 UpperCamelCase__ = [1.0] UpperCamelCase__ = [0.0] UpperCamelCase__ = 0.0 for i, qid in enumerate(__A ): if qid_to_has_ans[qid]: true_pos += scores[qid] UpperCamelCase__ = true_pos / float(i + 1 ) UpperCamelCase__ = true_pos / float(__A ) if i == len(__A ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(__A ) recalls.append(__A ) if out_image: plot_pr_curve(__A , __A , __A , __A ) return {"ap": 100.0 * avg_prec} def _UpperCamelCase ( __A , __A , __A , __A , __A , __A ) -> List[str]: '''simple docstring''' if out_image_dir and not os.path.exists(__A ): os.makedirs(__A ) UpperCamelCase__ = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return UpperCamelCase__ = make_precision_recall_eval( __A , __A , __A , __A , out_image=os.path.join(__A , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) UpperCamelCase__ = make_precision_recall_eval( __A , __A , __A , __A , out_image=os.path.join(__A , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) UpperCamelCase__ = {k: float(__A ) for k, v in qid_to_has_ans.items()} UpperCamelCase__ = make_precision_recall_eval( __A , __A , __A , __A , out_image=os.path.join(__A , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(__A , __A , "pr_exact" ) merge_eval(__A , __A , "pr_f1" ) merge_eval(__A , __A , "pr_oracle" ) def _UpperCamelCase ( __A , __A , __A , __A ) -> List[str]: '''simple docstring''' if not qid_list: return UpperCamelCase__ = [na_probs[k] for k in qid_list] UpperCamelCase__ = np.ones_like(__A ) / float(len(__A ) ) plt.hist(__A , weights=__A , bins=20 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(F'''Histogram of no-answer probability: {name}''' ) plt.savefig(os.path.join(__A , F'''na_prob_hist_{name}.png''' ) ) plt.clf() def _UpperCamelCase ( __A , __A , __A , __A ) -> Tuple: '''simple docstring''' UpperCamelCase__ = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) UpperCamelCase__ = num_no_ans UpperCamelCase__ = cur_score UpperCamelCase__ = 0.0 UpperCamelCase__ = sorted(__A , key=lambda __A : na_probs[k] ) for i, qid in enumerate(__A ): if qid not in scores: continue if qid_to_has_ans[qid]: UpperCamelCase__ = scores[qid] else: if preds[qid]: UpperCamelCase__ = -1 else: UpperCamelCase__ = 0 cur_score += diff if cur_score > best_score: UpperCamelCase__ = cur_score UpperCamelCase__ = na_probs[qid] return 100.0 * best_score / len(__A ), best_thresh def _UpperCamelCase ( __A , __A , __A , __A , __A , __A ) -> Dict: '''simple docstring''' UpperCamelCase__ , UpperCamelCase__ = find_best_thresh(__A , __A , __A , __A ) UpperCamelCase__ , UpperCamelCase__ = find_best_thresh(__A , __A , __A , __A ) UpperCamelCase__ = best_exact UpperCamelCase__ = exact_thresh UpperCamelCase__ = best_fa UpperCamelCase__ = fa_thresh def _UpperCamelCase ( ) -> Any: '''simple docstring''' with open(OPTS.data_file ) as f: UpperCamelCase__ = json.load(__A ) UpperCamelCase__ = dataset_json["data"] with open(OPTS.pred_file ) as f: UpperCamelCase__ = json.load(__A ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: UpperCamelCase__ = json.load(__A ) else: UpperCamelCase__ = {k: 0.0 for k in preds} UpperCamelCase__ = make_qid_to_has_ans(__A ) # maps qid to True/False UpperCamelCase__ = [k for k, v in qid_to_has_ans.items() if v] UpperCamelCase__ = [k for k, v in qid_to_has_ans.items() if not v] UpperCamelCase__ , UpperCamelCase__ = get_raw_scores(__A , __A ) UpperCamelCase__ = apply_no_ans_threshold(__A , __A , __A , OPTS.na_prob_thresh ) UpperCamelCase__ = apply_no_ans_threshold(__A , __A , __A , OPTS.na_prob_thresh ) UpperCamelCase__ = make_eval_dict(__A , __A ) if has_ans_qids: UpperCamelCase__ = make_eval_dict(__A , __A , qid_list=__A ) merge_eval(__A , __A , "HasAns" ) if no_ans_qids: UpperCamelCase__ = make_eval_dict(__A , __A , qid_list=__A ) merge_eval(__A , __A , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(__A , __A , __A , __A , __A , __A ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(__A , __A , __A , __A , __A , OPTS.out_image_dir ) histogram_na_prob(__A , __A , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(__A , __A , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(__A , __A ) else: print(json.dumps(__A , indent=2 ) ) if __name__ == "__main__": a__ : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('Agg') import matplotlib.pyplot as plt main()
80
0
from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __A ( lowerCAmelCase__ ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(a__ , 'embed_dim' ) ) self.parent.assertTrue(hasattr(a__ , 'num_heads' ) ) class __A : """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=13 , lowerCamelCase__=64 , lowerCamelCase__=3 , lowerCamelCase__=[16, 48, 96] , lowerCamelCase__=[1, 3, 6] , lowerCamelCase__=[1, 2, 10] , lowerCamelCase__=[7, 3, 3] , lowerCamelCase__=[4, 2, 2] , lowerCamelCase__=[2, 1, 1] , lowerCamelCase__=[2, 2, 2] , lowerCamelCase__=[False, False, True] , lowerCamelCase__=[0.0, 0.0, 0.0] , lowerCamelCase__=0.02 , lowerCamelCase__=1E-12 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=2 , ): """simple docstring""" __UpperCamelCase : Optional[Any] =parent __UpperCamelCase : int =batch_size __UpperCamelCase : Optional[int] =image_size __UpperCamelCase : Union[str, Any] =patch_sizes __UpperCamelCase : str =patch_stride __UpperCamelCase : Tuple =patch_padding __UpperCamelCase : Optional[Any] =is_training __UpperCamelCase : str =use_labels __UpperCamelCase : List[Any] =num_labels __UpperCamelCase : Optional[int] =num_channels __UpperCamelCase : List[str] =embed_dim __UpperCamelCase : Optional[int] =num_heads __UpperCamelCase : Dict =stride_kv __UpperCamelCase : Any =depth __UpperCamelCase : List[Any] =cls_token __UpperCamelCase : List[str] =attention_drop_rate __UpperCamelCase : List[Any] =initializer_range __UpperCamelCase : Tuple =layer_norm_eps def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCamelCase : Optional[int] =None if self.use_labels: # create a random int32 tensor of given shape __UpperCamelCase : int =ids_tensor([self.batch_size] , self.num_labels ) __UpperCamelCase : List[Any] =self.get_config() return config, pixel_values, labels def __lowercase ( self ): """simple docstring""" return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : int =TFCvtModel(config=a__ ) __UpperCamelCase : int =model(a__ , training=a__ ) __UpperCamelCase : Any =(self.image_size, self.image_size) __UpperCamelCase , __UpperCamelCase : int =image_size[0], image_size[1] for i in range(len(self.depth ) ): __UpperCamelCase : int =floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __UpperCamelCase : List[str] =floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.num_labels __UpperCamelCase : Optional[Any] =TFCvtForImageClassification(a__ ) __UpperCamelCase : Any =model(a__ , labels=a__ , training=a__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =self.prepare_config_and_inputs() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase : Tuple =config_and_inputs __UpperCamelCase : int ={'pixel_values': pixel_values} return config, inputs_dict @require_tf class __A ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : str =(TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () UpperCamelCase__ : Dict =( {"feature-extraction": TFCvtModel, "image-classification": TFCvtForImageClassification} if is_tf_available() else {} ) UpperCamelCase__ : int =False UpperCamelCase__ : Any =False UpperCamelCase__ : int =False UpperCamelCase__ : Any =False UpperCamelCase__ : int =False def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =TFCvtModelTester(self ) __UpperCamelCase : List[str] =TFCvtConfigTester(self , config_class=a__ , has_text_modality=a__ , hidden_size=37 ) def __lowercase ( self ): """simple docstring""" self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason='Cvt does not output attentions' ) def __lowercase ( self ): """simple docstring""" pass @unittest.skip(reason='Cvt does not use inputs_embeds' ) def __lowercase ( self ): """simple docstring""" pass @unittest.skip(reason='Cvt does not support input and output embeddings' ) def __lowercase ( self ): """simple docstring""" pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU' ) ) == 0 , reason='TF does not support backprop for grouped convolutions on CPU.' , ) def __lowercase ( self ): """simple docstring""" super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices('GPU' ) ) == 0 , reason='TF does not support backprop for grouped convolutions on CPU.' , ) @slow def __lowercase ( self ): """simple docstring""" super().test_keras_fit() @unittest.skip(reason='Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =tf.keras.mixed_precision.Policy('mixed_float16' ) tf.keras.mixed_precision.set_global_policy(a__ ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy('float32' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : List[Any] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase : Optional[Any] =model_class(a__ ) __UpperCamelCase : List[str] =inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCamelCase : Optional[Any] =[*signature.parameters.keys()] __UpperCamelCase : Optional[int] =['pixel_values'] self.assertListEqual(arg_names[:1] , a__ ) def __lowercase ( self ): """simple docstring""" def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): __UpperCamelCase : Optional[Any] =model_class(a__ ) __UpperCamelCase : Union[str, Any] =model(**self._prepare_for_class(a__ , a__ ) ) __UpperCamelCase : Optional[Any] =outputs.hidden_states __UpperCamelCase : Optional[int] =len(self.model_tester.depth ) self.assertEqual(len(a__ ) , a__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) __UpperCamelCase , __UpperCamelCase : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase : Optional[int] =True check_hidden_states_output(a__ , a__ , a__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCamelCase : Dict =True check_hidden_states_output(a__ , a__ , a__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : int =TFCvtModel.from_pretrained(a__ ) self.assertIsNotNone(a__ ) def A ( ) -> Optional[Any]: __UpperCamelCase : int =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_tf @require_vision class __A ( unittest.TestCase ): """simple docstring""" @cached_property def __lowercase ( self ): """simple docstring""" return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __UpperCamelCase : Optional[Any] =self.default_image_processor __UpperCamelCase : Dict =prepare_img() __UpperCamelCase : Tuple =image_processor(images=a__ , return_tensors='tf' ) # forward pass __UpperCamelCase : Optional[int] =model(**a__ ) # verify the logits __UpperCamelCase : Optional[Any] =tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , a__ ) __UpperCamelCase : Optional[int] =tf.constant([0.9_285, 0.9_015, -0.3_150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , a__ , atol=1E-4 ) )
364
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotSmallConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor if is_flax_available(): import os # The slow tests are often failing with OOM error on GPU # This makes JAX allocate exactly what is needed on demand, and deallocate memory that is no longer needed # but will be slower as stated here https://jax.readthedocs.io/en/latest/gpu_memory_allocation.html A_ :Tuple = '''platform''' import jax import jax.numpy as jnp from transformers.models.blenderbot_small.modeling_flax_blenderbot_small import ( FlaxBlenderbotSmallForConditionalGeneration, FlaxBlenderbotSmallModel, shift_tokens_right, ) def A ( a_ ,a_ ,a_=None ,a_=None ,a_=None ,a_=None ,a_=None ,a_=None ,) -> int: if attention_mask is None: __UpperCamelCase : Any =np.where(input_ids != config.pad_token_id ,1 ,0 ) if decoder_attention_mask is None: __UpperCamelCase : Optional[Any] =np.where(decoder_input_ids != config.pad_token_id ,1 ,0 ) if head_mask is None: __UpperCamelCase : str =np.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: __UpperCamelCase : Any =np.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: __UpperCamelCase : Optional[int] =np.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, } class __A : """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=13 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=99 , lowerCamelCase__=16 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=32 , lowerCamelCase__=2 , lowerCamelCase__=1 , lowerCamelCase__=0 , lowerCamelCase__=0.02 , ): """simple docstring""" __UpperCamelCase : Tuple =parent __UpperCamelCase : str =batch_size __UpperCamelCase : Optional[Any] =seq_length __UpperCamelCase : List[Any] =is_training __UpperCamelCase : int =use_labels __UpperCamelCase : int =vocab_size __UpperCamelCase : Any =hidden_size __UpperCamelCase : List[str] =num_hidden_layers __UpperCamelCase : Any =num_attention_heads __UpperCamelCase : int =intermediate_size __UpperCamelCase : List[Any] =hidden_act __UpperCamelCase : Optional[Any] =hidden_dropout_prob __UpperCamelCase : int =attention_probs_dropout_prob __UpperCamelCase : Tuple =max_position_embeddings __UpperCamelCase : List[Any] =eos_token_id __UpperCamelCase : Tuple =pad_token_id __UpperCamelCase : Any =bos_token_id __UpperCamelCase : Tuple =initializer_range def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =np.clip(ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) , 3 , self.vocab_size ) __UpperCamelCase : Any =np.concatenate((input_ids, 2 * np.ones((self.batch_size, 1) , dtype=np.intaa )) , -1 ) __UpperCamelCase : Any =shift_tokens_right(lowerCamelCase__ , 1 , 2 ) __UpperCamelCase : List[str] =BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , initializer_range=self.initializer_range , use_cache=lowerCamelCase__ , ) __UpperCamelCase : Dict =prepare_blenderbot_inputs_dict(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return config, inputs_dict def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Optional[Any] =self.prepare_config_and_inputs() return config, inputs_dict def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] =20 __UpperCamelCase : Optional[int] =model_class_name(lowerCamelCase__ ) __UpperCamelCase : Tuple =model.encode(inputs_dict['input_ids'] ) __UpperCamelCase , __UpperCamelCase : Optional[Any] =( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) __UpperCamelCase : Any =model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =jnp.ones((decoder_input_ids.shape[0], max_decoder_length) , dtype='i4' ) __UpperCamelCase : List[str] =jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) __UpperCamelCase : Any =model.decode( decoder_input_ids[:, :-1] , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , past_key_values=lowerCamelCase__ , decoder_position_ids=lowerCamelCase__ , ) __UpperCamelCase : List[str] =jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) __UpperCamelCase : Tuple =model.decode( decoder_input_ids[:, -1:] , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , past_key_values=outputs_cache.past_key_values , decoder_position_ids=lowerCamelCase__ , ) __UpperCamelCase : Tuple =model.decode(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Tuple =np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'Max diff is {diff}' ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" __UpperCamelCase : Optional[Any] =20 __UpperCamelCase : int =model_class_name(lowerCamelCase__ ) __UpperCamelCase : Optional[int] =model.encode(inputs_dict['input_ids'] ) __UpperCamelCase , __UpperCamelCase : Any =( inputs_dict['decoder_input_ids'], inputs_dict['decoder_attention_mask'], ) __UpperCamelCase : int =jnp.concatenate( [ decoder_attention_mask, jnp.zeros((decoder_attention_mask.shape[0], max_decoder_length - decoder_attention_mask.shape[1]) ), ] , axis=-1 , ) __UpperCamelCase : List[str] =model.init_cache(decoder_input_ids.shape[0] , lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Any =jnp.broadcast_to( jnp.arange(decoder_input_ids.shape[-1] - 1 )[None, :] , (decoder_input_ids.shape[0], decoder_input_ids.shape[-1] - 1) , ) __UpperCamelCase : Any =model.decode( decoder_input_ids[:, :-1] , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , past_key_values=lowerCamelCase__ , decoder_position_ids=lowerCamelCase__ , ) __UpperCamelCase : Optional[Any] =jnp.array(decoder_input_ids.shape[0] * [[decoder_input_ids.shape[-1] - 1]] , dtype='i4' ) __UpperCamelCase : Tuple =model.decode( decoder_input_ids[:, -1:] , lowerCamelCase__ , past_key_values=outputs_cache.past_key_values , decoder_attention_mask=lowerCamelCase__ , decoder_position_ids=lowerCamelCase__ , ) __UpperCamelCase : Optional[int] =model.decode(lowerCamelCase__ , lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ ) __UpperCamelCase : Tuple =np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f'Max diff is {diff}' ) @require_flax class __A ( unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Any =9_9 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Any =np.array( [ [71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 82, 2], [5, 97, 17, 39, 94, 40, 2], [76, 83, 94, 25, 70, 78, 2], [87, 59, 41, 35, 48, 66, 2], [55, 13, 16, 58, 5, 2, 1], # note padding [64, 27, 31, 51, 12, 75, 2], [52, 64, 86, 17, 83, 39, 2], [48, 61, 9, 24, 71, 82, 2], [26, 1, 60, 48, 22, 13, 2], [21, 5, 62, 28, 14, 76, 2], [45, 98, 37, 86, 59, 48, 2], [70, 70, 50, 9, 28, 0, 2], ] , dtype=np.intaa , ) __UpperCamelCase : int =input_ids.shape[0] __UpperCamelCase : List[str] =BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=24 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=32 , decoder_ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase , __UpperCamelCase : int =self._get_config_and_data() __UpperCamelCase : Optional[Any] =FlaxBlenderbotSmallForConditionalGeneration(lowerCamelCase__ ) __UpperCamelCase : int =lm_model(input_ids=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =(batch_size, input_ids.shape[1], config.vocab_size) self.assertEqual(outputs['logits'].shape , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =BlenderbotSmallConfig( vocab_size=self.vocab_size , d_model=14 , encoder_layers=2 , decoder_layers=2 , encoder_attention_heads=2 , decoder_attention_heads=2 , encoder_ffn_dim=8 , decoder_ffn_dim=8 , max_position_embeddings=48 , ) __UpperCamelCase : List[str] =FlaxBlenderbotSmallForConditionalGeneration(lowerCamelCase__ ) __UpperCamelCase : Any =np.array([[71, 82, 18, 33, 46, 91, 2], [68, 34, 26, 58, 30, 2, 1]] , dtype=np.intaa ) __UpperCamelCase : int =np.array([[82, 71, 82, 18, 2], [58, 68, 2, 1, 1]] , dtype=np.intaa ) __UpperCamelCase : Any =lm_model(input_ids=lowerCamelCase__ , decoder_input_ids=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =(*summary.shape, config.vocab_size) self.assertEqual(outputs['logits'].shape , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =np.array([[71, 82, 18, 33, 2, 1, 1], [68, 34, 26, 58, 30, 82, 2]] , dtype=np.intaa ) __UpperCamelCase : Tuple =shift_tokens_right(lowerCamelCase__ , 1 , 2 ) __UpperCamelCase : Optional[int] =np.equal(lowerCamelCase__ , 1 ).astype(np.floataa ).sum() __UpperCamelCase : Optional[Any] =np.equal(lowerCamelCase__ , 1 ).astype(np.floataa ).sum() self.assertEqual(shifted.shape , input_ids.shape ) self.assertEqual(lowerCamelCase__ , n_pad_before - 1 ) self.assertTrue(np.equal(shifted[:, 0] , 2 ).all() ) @require_flax class __A ( a , unittest.TestCase , a ): """simple docstring""" UpperCamelCase__ : Any =True UpperCamelCase__ : List[Any] =( ( FlaxBlenderbotSmallModel, FlaxBlenderbotSmallForConditionalGeneration, ) if is_flax_available() else () ) UpperCamelCase__ : Dict =(FlaxBlenderbotSmallForConditionalGeneration,) if is_flax_available() else () def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[int] =FlaxBlenderbotSmallModelTester(self ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Dict =self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Any =self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: self.model_tester.check_use_cache_forward_with_attn_mask(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __UpperCamelCase : Tuple =self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =model_class(lowerCamelCase__ ) @jax.jit def encode_jitted(lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): return model.encode(input_ids=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) with self.subTest('JIT Enabled' ): __UpperCamelCase : Union[str, Any] =encode_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): __UpperCamelCase : Any =encode_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase , __UpperCamelCase : Optional[int] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __UpperCamelCase : Tuple =model_class(lowerCamelCase__ ) __UpperCamelCase : Tuple =model.encode(inputs_dict['input_ids'] , inputs_dict['attention_mask'] ) __UpperCamelCase : Any ={ 'decoder_input_ids': inputs_dict['decoder_input_ids'], 'decoder_attention_mask': inputs_dict['decoder_attention_mask'], 'encoder_outputs': encoder_outputs, } @jax.jit def decode_jitted(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return model.decode( decoder_input_ids=lowerCamelCase__ , decoder_attention_mask=lowerCamelCase__ , encoder_outputs=lowerCamelCase__ , ) with self.subTest('JIT Enabled' ): __UpperCamelCase : Optional[Any] =decode_jitted(**lowerCamelCase__ ).to_tuple() with self.subTest('JIT Disabled' ): with jax.disable_jit(): __UpperCamelCase : int =decode_jitted(**lowerCamelCase__ ).to_tuple() self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) for jitted_output, output in zip(lowerCamelCase__ , lowerCamelCase__ ): self.assertEqual(jitted_output.shape , output.shape ) @slow def __lowercase ( self ): """simple docstring""" for model_class_name in self.all_model_classes: __UpperCamelCase : Optional[Any] =model_class_name.from_pretrained('facebook/blenderbot_small-90M' ) # FlaxBlenderbotForSequenceClassification expects eos token in input_ids __UpperCamelCase : Optional[Any] =np.ones((1, 1) ) * model.config.eos_token_id __UpperCamelCase : Union[str, Any] =model(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ )
245
0