code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : Optional[Any]=13 , lowerCamelCase : Optional[Any]=7 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : List[Any]=True , lowerCamelCase : List[str]=99 , lowerCamelCase : str=64 , lowerCamelCase : Tuple=5 , lowerCamelCase : str=4 , lowerCamelCase : Any=37 , lowerCamelCase : Union[str, Any]="gelu" , lowerCamelCase : str=0.1 , lowerCamelCase : str=0.1 , lowerCamelCase : Any=512 , lowerCamelCase : Tuple=16 , lowerCamelCase : List[str]=2 , lowerCamelCase : Dict=0.02 , lowerCamelCase : List[Any]=3 , lowerCamelCase : Any=4 , lowerCamelCase : Optional[int]=None , ) -> int: __snake_case : Union[str, Any] = parent __snake_case : List[str] = batch_size __snake_case : str = seq_length __snake_case : Tuple = is_training __snake_case : str = use_input_mask __snake_case : Any = use_token_type_ids __snake_case : List[Any] = use_labels __snake_case : str = vocab_size __snake_case : Any = hidden_size __snake_case : Optional[int] = num_hidden_layers __snake_case : Tuple = num_attention_heads __snake_case : Tuple = intermediate_size __snake_case : str = hidden_act __snake_case : List[Any] = hidden_dropout_prob __snake_case : Any = attention_probs_dropout_prob __snake_case : Dict = max_position_embeddings __snake_case : int = type_vocab_size __snake_case : Tuple = type_sequence_label_size __snake_case : Dict = initializer_range __snake_case : Tuple = num_labels __snake_case : str = num_choices __snake_case : Union[str, Any] = scope __snake_case : int = vocab_size - 1 def __snake_case ( self : Tuple ) -> Optional[int]: __snake_case : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : Any = None if self.use_input_mask: __snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case : Dict = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, input_ids, input_mask, token_labels def __snake_case ( self : Optional[Any] ) -> List[str]: return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def __snake_case ( self : List[str] ) -> str: __snake_case , __snake_case , __snake_case , __snake_case : int = self.prepare_config_and_inputs() __snake_case : List[Any] = True return config, input_ids, input_mask, token_labels def __snake_case ( self : Optional[Any] , lowerCamelCase : Optional[int] , lowerCamelCase : int , lowerCamelCase : List[str] ) -> Optional[Any]: __snake_case : Dict = GPTNeoXModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , attention_mask=lowerCamelCase ) __snake_case : Any = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self : str , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] , lowerCamelCase : Optional[int] ) -> int: __snake_case : Tuple = True __snake_case : Dict = GPTNeoXModel(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Union[str, Any] = model(lowerCamelCase , attention_mask=lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self : int , lowerCamelCase : Any , lowerCamelCase : List[Any] , lowerCamelCase : List[str] , lowerCamelCase : Optional[Any] ) -> Tuple: __snake_case : str = GPTNeoXForCausalLM(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Any = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case ( self : Optional[int] , lowerCamelCase : int , lowerCamelCase : Union[str, Any] , lowerCamelCase : Any , lowerCamelCase : Union[str, Any] ) -> Any: __snake_case : Any = self.num_labels __snake_case : str = GPTNeoXForQuestionAnswering(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Optional[Any] = model(lowerCamelCase , attention_mask=lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __snake_case ( self : Optional[Any] , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Dict , lowerCamelCase : Dict ) -> Union[str, Any]: __snake_case : Any = self.num_labels __snake_case : List[Any] = GPTNeoXForSequenceClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case : Union[str, Any] = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Optional[Any] , lowerCamelCase : str , lowerCamelCase : Any , lowerCamelCase : int , lowerCamelCase : Any ) -> Optional[int]: __snake_case : Tuple = self.num_labels __snake_case : Dict = GPTNeoXForTokenClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __snake_case ( self : Optional[Any] , lowerCamelCase : Tuple , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> List[Any]: __snake_case : Tuple = True __snake_case : int = GPTNeoXForCausalLM(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() # first forward pass __snake_case : Tuple = model(lowerCamelCase , attention_mask=lowerCamelCase , use_cache=lowerCamelCase ) __snake_case : Any = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case : Dict = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case : Dict = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case : Union[str, Any] = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case : Optional[int] = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case : str = model(lowerCamelCase , attention_mask=lowerCamelCase , output_hidden_states=lowerCamelCase ) __snake_case : Optional[int] = output_from_no_past["hidden_states"][0] __snake_case : Optional[Any] = model( lowerCamelCase , attention_mask=lowerCamelCase , past_key_values=lowerCamelCase , output_hidden_states=lowerCamelCase , )["hidden_states"][0] # select random slice __snake_case : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case : str = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case : Tuple = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-3 ) ) def __snake_case ( self : List[Any] ) -> Tuple: __snake_case : List[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Dict = config_and_inputs __snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : int = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) __UpperCAmelCase : str = (GPTNeoXForCausalLM,) if is_torch_available() else () __UpperCAmelCase : Any = ( { "feature-extraction": GPTNeoXModel, "question-answering": GPTNeoXForQuestionAnswering, "text-classification": GPTNeoXForSequenceClassification, "text-generation": GPTNeoXForCausalLM, "token-classification": GPTNeoXForTokenClassification, "zero-shot": GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) __UpperCAmelCase : List[Any] = False __UpperCAmelCase : List[str] = False __UpperCAmelCase : Dict = False __UpperCAmelCase : str = False def __snake_case ( self : Optional[int] ) -> int: __snake_case : List[str] = GPTNeoXModelTester(self ) __snake_case : str = ConfigTester(self , config_class=lowerCamelCase , hidden_size=64 , num_attention_heads=8 ) def __snake_case ( self : str ) -> Tuple: self.config_tester.run_common_tests() def __snake_case ( self : Optional[Any] ) -> str: __snake_case , __snake_case , __snake_case , __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> List[str]: __snake_case , __snake_case , __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : str ) -> Optional[int]: # This regression test was failing with PyTorch < 1.3 __snake_case , __snake_case , __snake_case , __snake_case : str = self.model_tester.prepare_config_and_inputs_for_decoder() __snake_case : Tuple = None self.model_tester.create_and_check_model_as_decoder(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case , __snake_case , __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : str ) -> Dict: __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*lowerCamelCase ) def __snake_case ( self : str ) -> Dict: __snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Tuple: __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase ) @unittest.skip(reason="Feed forward chunking is not implemented" ) def __snake_case ( self : List[Any] ) -> Optional[int]: pass @parameterized.expand([("linear",), ("dynamic",)] ) def __snake_case ( self : str , lowerCamelCase : Optional[int] ) -> str: __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : int = ids_tensor([1, 10] , config.vocab_size ) __snake_case : Tuple = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __snake_case : str = GPTNeoXModel(lowerCamelCase ) original_model.to(lowerCamelCase ) original_model.eval() __snake_case : List[Any] = original_model(lowerCamelCase ).last_hidden_state __snake_case : Tuple = original_model(lowerCamelCase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights __snake_case : Union[str, Any] = {"type": scaling_type, "factor": 10.0} __snake_case : List[Any] = GPTNeoXModel(lowerCamelCase ) scaled_model.to(lowerCamelCase ) scaled_model.eval() __snake_case : List[Any] = scaled_model(lowerCamelCase ).last_hidden_state __snake_case : Any = scaled_model(lowerCamelCase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-5 ) ) @require_torch class a (unittest.TestCase ): """simple docstring""" @slow def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : List[Any] = AutoTokenizer.from_pretrained("EleutherAI/pythia-410m-deduped" ) for checkpointing in [True, False]: __snake_case : Union[str, Any] = GPTNeoXForCausalLM.from_pretrained("EleutherAI/pythia-410m-deduped" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(lowerCamelCase ) __snake_case : Dict = tokenizer("My favorite food is" , return_tensors="pt" ).to(lowerCamelCase ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 __snake_case : Optional[Any] = "My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure" __snake_case : List[Any] = model.generate(**lowerCamelCase , do_sample=lowerCamelCase , max_new_tokens=20 ) __snake_case : List[str] = tokenizer.batch_decode(lowerCamelCase )[0] self.assertEqual(lowerCamelCase , lowerCamelCase )
81
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
1
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class a (unittest.TestCase ): """simple docstring""" @slow def __snake_case ( self : int ) -> Optional[Any]: __snake_case : List[str] = TFCamembertModel.from_pretrained("jplu/tf-camembert-base" ) __snake_case : str = tf.convert_to_tensor( [[5, 121, 11, 660, 16, 730, 25543, 110, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" __snake_case : str = model(lowerCamelCase )["last_hidden_state"] __snake_case : Union[str, Any] = tf.TensorShape((1, 10, 768) ) self.assertEqual(output.shape , lowerCamelCase ) # compare the actual values for a slice. __snake_case : Tuple = tf.convert_to_tensor( [[[-0.02_54, 0.02_35, 0.10_27], [0.06_06, -0.18_11, -0.04_18], [-0.15_61, -0.11_27, 0.26_87]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
81
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
1
import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _snake_case : str = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Dict = "AutoTokenizer" __UpperCAmelCase : List[Any] = ["tokenizer"] __UpperCAmelCase : Tuple = { "semantic_prompt": 1, "coarse_prompt": 2, "fine_prompt": 2, } def __init__( self : str , lowerCamelCase : Tuple , lowerCamelCase : Union[str, Any]=None ) -> Optional[Any]: super().__init__(lowerCamelCase ) __snake_case : Tuple = speaker_embeddings @classmethod def __snake_case ( cls : Optional[Any] , lowerCamelCase : int , lowerCamelCase : Optional[int]="speaker_embeddings_path.json" , **lowerCamelCase : List[Any] ) -> List[str]: if speaker_embeddings_dict_path is not None: __snake_case : Any = get_file_from_repo( lowerCamelCase , lowerCamelCase , subfolder=kwargs.pop("subfolder" , lowerCamelCase ) , cache_dir=kwargs.pop("cache_dir" , lowerCamelCase ) , force_download=kwargs.pop("force_download" , lowerCamelCase ) , proxies=kwargs.pop("proxies" , lowerCamelCase ) , resume_download=kwargs.pop("resume_download" , lowerCamelCase ) , local_files_only=kwargs.pop("local_files_only" , lowerCamelCase ) , use_auth_token=kwargs.pop("use_auth_token" , lowerCamelCase ) , revision=kwargs.pop("revision" , lowerCamelCase ) , ) if speaker_embeddings_path is None: logger.warning( F'`{os.path.join(lowerCamelCase , lowerCamelCase )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.' ) __snake_case : List[str] = None else: with open(lowerCamelCase ) as speaker_embeddings_json: __snake_case : List[Any] = json.load(lowerCamelCase ) else: __snake_case : Dict = None __snake_case : Union[str, Any] = AutoTokenizer.from_pretrained(lowerCamelCase , **lowerCamelCase ) return cls(tokenizer=lowerCamelCase , speaker_embeddings=lowerCamelCase ) def __snake_case ( self : Optional[int] , lowerCamelCase : Any , lowerCamelCase : List[str]="speaker_embeddings_path.json" , lowerCamelCase : List[Any]="speaker_embeddings" , lowerCamelCase : bool = False , **lowerCamelCase : List[Any] , ) -> str: if self.speaker_embeddings is not None: os.makedirs(os.path.join(lowerCamelCase , lowerCamelCase , "v2" ) , exist_ok=lowerCamelCase ) __snake_case : List[Any] = {} __snake_case : Optional[int] = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": __snake_case : List[str] = self._load_voice_preset(lowerCamelCase ) __snake_case : int = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , lowerCamelCase , F'{prompt_key}_{key}' ) , voice_preset[key] , allow_pickle=lowerCamelCase , ) __snake_case : Optional[Any] = os.path.join(lowerCamelCase , F'{prompt_key}_{key}.npy' ) __snake_case : List[str] = tmp_dict with open(os.path.join(lowerCamelCase , lowerCamelCase ) , "w" ) as fp: json.dump(lowerCamelCase , lowerCamelCase ) super().save_pretrained(lowerCamelCase , lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : str = None , **lowerCamelCase : int ) -> Union[str, Any]: __snake_case : Optional[int] = self.speaker_embeddings[voice_preset] __snake_case : List[str] = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].' ) __snake_case : int = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , lowerCamelCase ) , cache_dir=kwargs.pop("cache_dir" , lowerCamelCase ) , force_download=kwargs.pop("force_download" , lowerCamelCase ) , proxies=kwargs.pop("proxies" , lowerCamelCase ) , resume_download=kwargs.pop("resume_download" , lowerCamelCase ) , local_files_only=kwargs.pop("local_files_only" , lowerCamelCase ) , use_auth_token=kwargs.pop("use_auth_token" , lowerCamelCase ) , revision=kwargs.pop("revision" , lowerCamelCase ) , ) if path is None: raise ValueError( F'`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.' ) __snake_case : Tuple = np.load(lowerCamelCase ) return voice_preset_dict def __snake_case ( self : Optional[int] , lowerCamelCase : Optional[dict] = None ) -> Tuple: for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'Voice preset unrecognized, missing {key} as a key.' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) def __call__( self : Optional[Any] , lowerCamelCase : Optional[Any]=None , lowerCamelCase : List[Any]=None , lowerCamelCase : Dict="pt" , lowerCamelCase : List[Any]=256 , lowerCamelCase : str=False , lowerCamelCase : Optional[int]=True , lowerCamelCase : str=False , **lowerCamelCase : Any , ) -> List[Any]: if voice_preset is not None and not isinstance(lowerCamelCase , lowerCamelCase ): if ( isinstance(lowerCamelCase , lowerCamelCase ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): __snake_case : int = self._load_voice_preset(lowerCamelCase ) else: if isinstance(lowerCamelCase , lowerCamelCase ) and not voice_preset.endswith(".npz" ): __snake_case : List[str] = voice_preset + ".npz" __snake_case : Union[str, Any] = np.load(lowerCamelCase ) if voice_preset is not None: self._validate_voice_preset_dict(lowerCamelCase , **lowerCamelCase ) __snake_case : str = BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase ) __snake_case : Dict = self.tokenizer( lowerCamelCase , return_tensors=lowerCamelCase , padding="max_length" , max_length=lowerCamelCase , return_attention_mask=lowerCamelCase , return_token_type_ids=lowerCamelCase , add_special_tokens=lowerCamelCase , **lowerCamelCase , ) if voice_preset is not None: __snake_case : Tuple = voice_preset return encoded_text
81
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
1
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sentencepiece class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> Optional[int]: __snake_case : List[str] = "laion/clap-htsat-unfused" __snake_case : int = tempfile.mkdtemp() def __snake_case ( self : List[Any] , **lowerCamelCase : List[str] ) -> Optional[int]: return RobertaTokenizer.from_pretrained(self.checkpoint , **lowerCamelCase ) def __snake_case ( self : Union[str, Any] , **lowerCamelCase : Tuple ) -> Optional[Any]: return ClapFeatureExtractor.from_pretrained(self.checkpoint , **lowerCamelCase ) def __snake_case ( self : Any ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Union[str, Any] ) -> str: __snake_case : List[Any] = self.get_tokenizer() __snake_case : Any = self.get_feature_extractor() __snake_case : str = ClapProcessor(tokenizer=lowerCamelCase , feature_extractor=lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) __snake_case : List[Any] = ClapProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowerCamelCase ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Union[str, Any]: __snake_case : List[str] = ClapProcessor(tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Union[str, Any] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) __snake_case : Union[str, Any] = self.get_feature_extractor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Any = ClapProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowerCamelCase ) self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.feature_extractor , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = self.get_feature_extractor() __snake_case : Dict = self.get_tokenizer() __snake_case : Any = ClapProcessor(tokenizer=lowerCamelCase , feature_extractor=lowerCamelCase ) __snake_case : List[str] = floats_list((3, 1000) ) __snake_case : List[str] = feature_extractor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(audios=lowerCamelCase , return_tensors="np" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __snake_case ( self : Any ) -> Optional[Any]: __snake_case : Dict = self.get_feature_extractor() __snake_case : Optional[Any] = self.get_tokenizer() __snake_case : str = ClapProcessor(tokenizer=lowerCamelCase , feature_extractor=lowerCamelCase ) __snake_case : Dict = "This is a test string" __snake_case : List[Any] = processor(text=lowerCamelCase ) __snake_case : Dict = tokenizer(lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : Dict = self.get_feature_extractor() __snake_case : Union[str, Any] = self.get_tokenizer() __snake_case : int = ClapProcessor(tokenizer=lowerCamelCase , feature_extractor=lowerCamelCase ) __snake_case : Optional[Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __snake_case : Tuple = processor.batch_decode(lowerCamelCase ) __snake_case : int = tokenizer.batch_decode(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> int: __snake_case : Dict = self.get_feature_extractor() __snake_case : Tuple = self.get_tokenizer() __snake_case : List[str] = ClapProcessor(tokenizer=lowerCamelCase , feature_extractor=lowerCamelCase ) self.assertListEqual( processor.model_input_names[2:] , feature_extractor.model_input_names , msg="`processor` and `feature_extractor` model input names do not match" , )
81
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Union[str, Any] = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", "InstructBlipVisionConfig", ], "processing_instructblip": ["InstructBlipProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ "INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "InstructBlipQFormerModel", "InstructBlipPreTrainedModel", "InstructBlipForConditionalGeneration", "InstructBlipVisionModel", ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
81
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
1
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class a : """simple docstring""" def __init__( self : Optional[Any] , lowerCamelCase : Tuple , lowerCamelCase : str=3 , lowerCamelCase : Tuple=32 , lowerCamelCase : Tuple=3 , lowerCamelCase : Any=10 , lowerCamelCase : int=[8, 16, 32, 64] , lowerCamelCase : Optional[int]=[1, 1, 2, 1] , lowerCamelCase : int=True , lowerCamelCase : Optional[Any]=True , lowerCamelCase : Any="relu" , lowerCamelCase : Tuple=3 , lowerCamelCase : str=None , lowerCamelCase : Optional[Any]=["stage2", "stage3", "stage4"] , lowerCamelCase : Union[str, Any]=[2, 3, 4] , lowerCamelCase : str=1 , ) -> Union[str, Any]: __snake_case : Optional[int] = parent __snake_case : Optional[int] = batch_size __snake_case : Any = image_size __snake_case : Optional[int] = num_channels __snake_case : int = embeddings_size __snake_case : Tuple = hidden_sizes __snake_case : Optional[int] = depths __snake_case : Union[str, Any] = is_training __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = hidden_act __snake_case : List[str] = num_labels __snake_case : Union[str, Any] = scope __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : int = out_features __snake_case : Tuple = out_indices __snake_case : int = num_groups def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Tuple = None if self.use_labels: __snake_case : List[Any] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : str = self.get_config() return config, pixel_values, labels def __snake_case ( self : int ) -> List[str]: return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : Dict , lowerCamelCase : str , lowerCamelCase : List[Any] ) -> Any: __snake_case : int = BitModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[int] , lowerCamelCase : Any , lowerCamelCase : Optional[int] ) -> int: __snake_case : List[str] = self.num_labels __snake_case : Union[str, Any] = BitForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Any = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : str , lowerCamelCase : Optional[Any] , lowerCamelCase : int ) -> str: __snake_case : List[Any] = BitBackbone(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Union[str, Any] = model(lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __snake_case : Dict = None __snake_case : str = BitBackbone(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : int = model(lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : Tuple = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case : Optional[Any] = config_and_inputs __snake_case : int = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () __UpperCAmelCase : List[str] = ( {"feature-extraction": BitModel, "image-classification": BitForImageClassification} if is_torch_available() else {} ) __UpperCAmelCase : int = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Any = False __UpperCAmelCase : List[str] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : int ) -> Optional[Any]: __snake_case : Any = BitModelTester(self ) __snake_case : Optional[Any] = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : str ) -> int: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __snake_case ( self : Optional[int] ) -> int: return @unittest.skip(reason="Bit does not output attentions" ) def __snake_case ( self : int ) -> Union[str, Any]: pass @unittest.skip(reason="Bit does not use inputs_embeds" ) def __snake_case ( self : int ) -> Dict: pass @unittest.skip(reason="Bit does not support input and output embeddings" ) def __snake_case ( self : str ) -> Optional[Any]: pass def __snake_case ( self : List[Any] ) -> List[Any]: __snake_case , __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Tuple = model_class(lowerCamelCase ) __snake_case : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Dict = [*signature.parameters.keys()] __snake_case : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : str ) -> List[str]: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : str ) -> Any: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Tuple: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Union[str, Any] = model_class(config=lowerCamelCase ) for name, module in model.named_modules(): if isinstance(lowerCamelCase , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) def __snake_case ( self : Dict ) -> List[Any]: def check_hidden_states_output(lowerCamelCase : Dict , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : Union[str, Any] = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : List[str] = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __snake_case : List[Any] = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __snake_case , __snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Tuple = ["preactivation", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: __snake_case : Optional[Any] = layer_type __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) @unittest.skip(reason="Bit does not use feedforward chunking" ) def __snake_case ( self : str ) -> Union[str, Any]: pass def __snake_case ( self : Dict ) -> str: __snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def __snake_case ( self : List[Any] ) -> List[Any]: for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Tuple = BitModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : Any ) -> Tuple: return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : str = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase ) __snake_case : str = self.default_image_processor __snake_case : Dict = prepare_img() __snake_case : Tuple = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Optional[int] = model(**lowerCamelCase ) # verify the logits __snake_case : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([[-0.65_26, -0.52_63, -1.43_98]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @require_torch class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = (BitBackbone,) if is_torch_available() else () __UpperCAmelCase : str = BitConfig __UpperCAmelCase : List[Any] = False def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[int] = BitModelTester(self )
81
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
1
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
1
from math import factorial class a : """simple docstring""" def __init__( self : Any , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> Any: __snake_case : Optional[Any] = real if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : List[str] = [1] * rank else: __snake_case : List[str] = rank def __repr__( self : Optional[int] ) -> List[str]: return ( F'{self.real}+' F'{"+".join(str(lowerCamelCase )+"E"+str(n+1 )for n,dual in enumerate(self.duals ) )}' ) def __snake_case ( self : Any ) -> Dict: __snake_case : Optional[int] = self.duals.copy() while cur[-1] == 0: cur.pop(-1 ) return Dual(self.real , lowerCamelCase ) def __add__( self : Any , lowerCamelCase : int ) -> Dict: if not isinstance(lowerCamelCase , lowerCamelCase ): return Dual(self.real + other , self.duals ) __snake_case : Dict = self.duals.copy() __snake_case : Tuple = other.duals.copy() if len(lowerCamelCase ) > len(lowerCamelCase ): o_dual.extend([1] * (len(lowerCamelCase ) - len(lowerCamelCase )) ) elif len(lowerCamelCase ) < len(lowerCamelCase ): s_dual.extend([1] * (len(lowerCamelCase ) - len(lowerCamelCase )) ) __snake_case : int = [] for i in range(len(lowerCamelCase ) ): new_duals.append(s_dual[i] + o_dual[i] ) return Dual(self.real + other.real , lowerCamelCase ) __UpperCAmelCase : Tuple = __add__ def __sub__( self : Union[str, Any] , lowerCamelCase : int ) -> List[Any]: return self + other * -1 def __mul__( self : Any , lowerCamelCase : Optional[Any] ) -> Any: if not isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Tuple = [] for i in self.duals: new_duals.append(i * other ) return Dual(self.real * other , lowerCamelCase ) __snake_case : Union[str, Any] = [0] * (len(self.duals ) + len(other.duals ) + 1) for i, item in enumerate(self.duals ): for j, jtem in enumerate(other.duals ): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals ) ): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals ) ): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real , lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = __mul__ def __truediv__( self : int , lowerCamelCase : str ) -> Optional[Any]: if not isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Optional[Any] = [] for i in self.duals: new_duals.append(i / other ) return Dual(self.real / other , lowerCamelCase ) raise ValueError def __floordiv__( self : Tuple , lowerCamelCase : Any ) -> Optional[Any]: if not isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : str = [] for i in self.duals: new_duals.append(i // other ) return Dual(self.real // other , lowerCamelCase ) raise ValueError def __pow__( self : Optional[int] , lowerCamelCase : List[Any] ) -> int: if n < 0 or isinstance(lowerCamelCase , lowerCamelCase ): raise ValueError("power must be a positive integer" ) if n == 0: return 1 if n == 1: return self __snake_case : Dict = self for _ in range(n - 1 ): x *= self return x def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if not callable(__lowerCamelCase ): raise ValueError("differentiate() requires a function as input for func" ) if not isinstance(__lowerCamelCase , (float, int) ): raise ValueError("differentiate() requires a float as input for position" ) if not isinstance(__lowerCamelCase , __lowerCamelCase ): raise ValueError("differentiate() requires an int as input for order" ) __snake_case : Optional[Any] = Dual(__lowerCamelCase , 1 ) __snake_case : Optional[int] = func(__lowerCamelCase ) if order == 0: return result.real return result.duals[order - 1] * factorial(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() def lowerCAmelCase_ ( __lowerCamelCase ): return y**2 * y**4 print(differentiate(f, 9, 2))
81
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
1
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging _snake_case : str = logging.get_logger(__name__) _snake_case : Tuple = { "Salesforce/codegen-350M-nl": "https://huggingface.co/Salesforce/codegen-350M-nl/resolve/main/config.json", "Salesforce/codegen-350M-multi": "https://huggingface.co/Salesforce/codegen-350M-multi/resolve/main/config.json", "Salesforce/codegen-350M-mono": "https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/config.json", "Salesforce/codegen-2B-nl": "https://huggingface.co/Salesforce/codegen-2B-nl/resolve/main/config.json", "Salesforce/codegen-2B-multi": "https://huggingface.co/Salesforce/codegen-2B-multi/resolve/main/config.json", "Salesforce/codegen-2B-mono": "https://huggingface.co/Salesforce/codegen-2B-mono/resolve/main/config.json", "Salesforce/codegen-6B-nl": "https://huggingface.co/Salesforce/codegen-6B-nl/resolve/main/config.json", "Salesforce/codegen-6B-multi": "https://huggingface.co/Salesforce/codegen-6B-multi/resolve/main/config.json", "Salesforce/codegen-6B-mono": "https://huggingface.co/Salesforce/codegen-6B-mono/resolve/main/config.json", "Salesforce/codegen-16B-nl": "https://huggingface.co/Salesforce/codegen-16B-nl/resolve/main/config.json", "Salesforce/codegen-16B-multi": "https://huggingface.co/Salesforce/codegen-16B-multi/resolve/main/config.json", "Salesforce/codegen-16B-mono": "https://huggingface.co/Salesforce/codegen-16B-mono/resolve/main/config.json", } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = "codegen" __UpperCAmelCase : Dict = { "max_position_embeddings": "n_positions", "hidden_size": "n_embd", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self : Optional[int] , lowerCamelCase : Dict=50400 , lowerCamelCase : int=2048 , lowerCamelCase : List[Any]=2048 , lowerCamelCase : List[Any]=4096 , lowerCamelCase : Dict=28 , lowerCamelCase : List[str]=16 , lowerCamelCase : str=64 , lowerCamelCase : Any=None , lowerCamelCase : Optional[int]="gelu_new" , lowerCamelCase : Any=0.0 , lowerCamelCase : Union[str, Any]=0.0 , lowerCamelCase : int=0.0 , lowerCamelCase : List[str]=1E-5 , lowerCamelCase : List[Any]=0.02 , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[Any]=50256 , lowerCamelCase : int=50256 , lowerCamelCase : str=False , **lowerCamelCase : int , ) -> Union[str, Any]: __snake_case : str = vocab_size __snake_case : List[Any] = n_ctx __snake_case : Tuple = n_positions __snake_case : int = n_embd __snake_case : List[str] = n_layer __snake_case : int = n_head __snake_case : Union[str, Any] = n_inner __snake_case : List[Any] = rotary_dim __snake_case : Tuple = activation_function __snake_case : str = resid_pdrop __snake_case : Dict = embd_pdrop __snake_case : Optional[int] = attn_pdrop __snake_case : Dict = layer_norm_epsilon __snake_case : Tuple = initializer_range __snake_case : Union[str, Any] = use_cache __snake_case : Dict = bos_token_id __snake_case : Any = eos_token_id super().__init__( bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase , tie_word_embeddings=lowerCamelCase , **lowerCamelCase ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : str , lowerCamelCase : PretrainedConfig , lowerCamelCase : str = "default" , lowerCamelCase : List[PatchingSpec] = None , lowerCamelCase : bool = False , ) -> Any: super().__init__(lowerCamelCase , task=lowerCamelCase , patching_specs=lowerCamelCase , use_past=lowerCamelCase ) if not getattr(self._config , "pad_token_id" , lowerCamelCase ): # TODO: how to do that better? __snake_case : Any = 0 @property def __snake_case ( self : Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: __snake_case : Dict = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(lowerCamelCase , direction="inputs" ) __snake_case : Optional[Any] = {0: "batch", 1: "past_sequence + sequence"} else: __snake_case : int = {0: "batch", 1: "sequence"} return common_inputs @property def __snake_case ( self : int ) -> int: return self._config.n_layer @property def __snake_case ( self : List[str] ) -> int: return self._config.n_head def __snake_case ( self : Tuple , lowerCamelCase : PreTrainedTokenizer , lowerCamelCase : int = -1 , lowerCamelCase : int = -1 , lowerCamelCase : bool = False , lowerCamelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: __snake_case : Union[str, Any] = super(lowerCamelCase , self ).generate_dummy_inputs( lowerCamelCase , batch_size=lowerCamelCase , seq_length=lowerCamelCase , is_pair=lowerCamelCase , framework=lowerCamelCase ) # We need to order the input in the way they appears in the forward() __snake_case : List[str] = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch __snake_case , __snake_case : Optional[Any] = common_inputs["input_ids"].shape # Not using the same length for past_key_values __snake_case : Optional[Any] = seqlen + 2 __snake_case : str = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) __snake_case : str = [ (torch.zeros(lowerCamelCase ), torch.zeros(lowerCamelCase )) for _ in range(self.num_layers ) ] __snake_case : List[Any] = common_inputs["attention_mask"] if self.use_past: __snake_case : List[Any] = ordered_inputs["attention_mask"].dtype __snake_case : Any = torch.cat( [ordered_inputs["attention_mask"], torch.ones(lowerCamelCase , lowerCamelCase , dtype=lowerCamelCase )] , dim=1 ) return ordered_inputs @property def __snake_case ( self : Tuple ) -> int: return 13
81
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
1
from importlib import import_module from .logging import get_logger _snake_case : Tuple = get_logger(__name__) class a : """simple docstring""" def __init__( self : str , lowerCamelCase : str , lowerCamelCase : Dict=None ) -> int: __snake_case : Dict = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("__" ): setattr(self , lowerCamelCase , getattr(lowerCamelCase , lowerCamelCase ) ) __snake_case : List[Any] = module._original_module if isinstance(lowerCamelCase , _PatchedModuleObj ) else module class a : """simple docstring""" __UpperCAmelCase : str = [] def __init__( self : Dict , lowerCamelCase : str , lowerCamelCase : str , lowerCamelCase : List[str] , lowerCamelCase : Any=None ) -> Optional[Any]: __snake_case : str = obj __snake_case : str = target __snake_case : Dict = new __snake_case : Union[str, Any] = target.split("." )[0] __snake_case : Any = {} __snake_case : str = attrs or [] def __enter__( self : Dict ) -> Optional[int]: *__snake_case , __snake_case : Union[str, Any] = self.target.split("." ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(lowerCamelCase ) ): try: __snake_case : int = import_module(".".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): __snake_case : Tuple = getattr(self.obj , lowerCamelCase ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(lowerCamelCase , _PatchedModuleObj ) and obj_attr._original_module is submodule) ): __snake_case : Optional[Any] = obj_attr # patch at top level setattr(self.obj , lowerCamelCase , _PatchedModuleObj(lowerCamelCase , attrs=self.attrs ) ) __snake_case : Union[str, Any] = getattr(self.obj , lowerCamelCase ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(lowerCamelCase , lowerCamelCase , _PatchedModuleObj(getattr(lowerCamelCase , lowerCamelCase , lowerCamelCase ) , attrs=self.attrs ) ) __snake_case : Dict = getattr(lowerCamelCase , lowerCamelCase ) # finally set the target attribute setattr(lowerCamelCase , lowerCamelCase , self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: __snake_case : List[Any] = getattr(import_module(".".join(lowerCamelCase ) ) , lowerCamelCase ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj , lowerCamelCase ) is attr_value: __snake_case : Dict = getattr(self.obj , lowerCamelCase ) setattr(self.obj , lowerCamelCase , self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" __snake_case : Tuple = globals()["__builtins__"][target_attr] setattr(self.obj , lowerCamelCase , self.new ) else: raise RuntimeError(F'Tried to patch attribute {target_attr} instead of a submodule.' ) def __exit__( self : Optional[int] , *lowerCamelCase : Optional[Any] ) -> Dict: for attr in list(self.original ): setattr(self.obj , lowerCamelCase , self.original.pop(lowerCamelCase ) ) def __snake_case ( self : Any ) -> Optional[int]: self.__enter__() self._active_patches.append(self ) def __snake_case ( self : List[str] ) -> Tuple: try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
81
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
1
import math import qiskit def lowerCAmelCase_ ( __lowerCamelCase = 1 , __lowerCamelCase = 1 , __lowerCamelCase = 1 ): if ( isinstance(__lowerCamelCase , __lowerCamelCase ) or isinstance(__lowerCamelCase , __lowerCamelCase ) or isinstance(__lowerCamelCase , __lowerCamelCase ) ): raise TypeError("inputs must be integers." ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("inputs must be positive." ) if ( (math.floor(__lowerCamelCase ) != input_a) or (math.floor(__lowerCamelCase ) != input_a) or (math.floor(__lowerCamelCase ) != carry_in) ): raise ValueError("inputs must be exact integers." ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("inputs must be less or equal to 2." ) # build registers __snake_case : int = qiskit.QuantumRegister(4 , "qr" ) __snake_case : Optional[int] = qiskit.ClassicalRegister(2 , "cr" ) # list the entries __snake_case : List[Any] = [input_a, input_a, carry_in] __snake_case : Dict = qiskit.QuantumCircuit(__lowerCamelCase , __lowerCamelCase ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(__lowerCamelCase ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(__lowerCamelCase ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(__lowerCamelCase ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , __lowerCamelCase ) # measure the last two qbits __snake_case : Optional[int] = qiskit.Aer.get_backend("aer_simulator" ) __snake_case : int = qiskit.execute(__lowerCamelCase , __lowerCamelCase , shots=1_0_0_0 ) return job.result().get_counts(__lowerCamelCase ) if __name__ == "__main__": print(f'''Total sum count for state is: {quantum_full_adder(1, 1, 1)}''')
81
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
1
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Dict = StableDiffusionInpaintPipeline __UpperCAmelCase : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS __UpperCAmelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS __UpperCAmelCase : int = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __UpperCAmelCase : List[Any] = frozenset([] ) def __snake_case ( self : List[Any] ) -> Tuple: torch.manual_seed(0 ) __snake_case : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase , ) __snake_case : Any = PNDMScheduler(skip_prk_steps=lowerCamelCase ) torch.manual_seed(0 ) __snake_case : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) __snake_case : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="gelu" , projection_dim=512 , ) __snake_case : Optional[Any] = CLIPTextModel(lowerCamelCase ) __snake_case : Optional[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) __snake_case : int = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def __snake_case ( self : List[Any] , lowerCamelCase : Optional[int] , lowerCamelCase : str=0 ) -> Any: # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched __snake_case : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase ) ).to(lowerCamelCase ) __snake_case : str = image.cpu().permute(0 , 2 , 3 , 1 )[0] __snake_case : Dict = Image.fromarray(np.uinta(lowerCamelCase ) ).convert("RGB" ).resize((64, 64) ) __snake_case : str = Image.fromarray(np.uinta(image + 4 ) ).convert("RGB" ).resize((64, 64) ) if str(lowerCamelCase ).startswith("mps" ): __snake_case : List[Any] = torch.manual_seed(lowerCamelCase ) else: __snake_case : Dict = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __snake_case : Any = { "prompt": "A painting of a squirrel eating a burger", "image": init_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : List[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator __snake_case : List[Any] = self.get_dummy_components() __snake_case : List[Any] = StableDiffusionInpaintPipeline(**lowerCamelCase ) __snake_case : List[Any] = sd_pipe.to(lowerCamelCase ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase ) __snake_case : int = self.get_dummy_inputs(lowerCamelCase ) __snake_case : Union[str, Any] = sd_pipe(**lowerCamelCase ).images __snake_case : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __snake_case : Dict = np.array([0.47_27, 0.57_35, 0.39_41, 0.54_46, 0.59_26, 0.43_94, 0.50_62, 0.46_54, 0.44_76] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __snake_case ( self : Dict ) -> Any: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Union[str, Any] ) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Any ) -> Tuple: __snake_case : Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) __snake_case : List[str] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) __snake_case : List[str] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench.npy" ) __snake_case : int = "stabilityai/stable-diffusion-2-inpainting" __snake_case : Tuple = StableDiffusionInpaintPipeline.from_pretrained(lowerCamelCase , safety_checker=lowerCamelCase ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing() __snake_case : List[Any] = "Face of a yellow cat, high resolution, sitting on a park bench" __snake_case : int = torch.manual_seed(0 ) __snake_case : Optional[Any] = pipe( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , generator=lowerCamelCase , output_type="np" , ) __snake_case : List[Any] = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9E-3 def __snake_case ( self : Dict ) -> Any: __snake_case : Union[str, Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) __snake_case : str = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) __snake_case : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench_fp16.npy" ) __snake_case : Any = "stabilityai/stable-diffusion-2-inpainting" __snake_case : Dict = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase , torch_dtype=torch.floataa , safety_checker=lowerCamelCase , ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing() __snake_case : str = "Face of a yellow cat, high resolution, sitting on a park bench" __snake_case : str = torch.manual_seed(0 ) __snake_case : Optional[int] = pipe( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , generator=lowerCamelCase , output_type="np" , ) __snake_case : Tuple = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5E-1 def __snake_case ( self : str ) -> Optional[Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __snake_case : List[str] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) __snake_case : str = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) __snake_case : Dict = "stabilityai/stable-diffusion-2-inpainting" __snake_case : Union[str, Any] = PNDMScheduler.from_pretrained(lowerCamelCase , subfolder="scheduler" ) __snake_case : Tuple = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase , safety_checker=lowerCamelCase , scheduler=lowerCamelCase , torch_dtype=torch.floataa , ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __snake_case : str = "Face of a yellow cat, high resolution, sitting on a park bench" __snake_case : List[str] = torch.manual_seed(0 ) __snake_case : str = pipe( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , generator=lowerCamelCase , num_inference_steps=2 , output_type="np" , ) __snake_case : int = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
81
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
1
def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(__lowerCamelCase ) ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): # Base Case if index == len(__lowerCamelCase ): return True # Recursive Step for i in range(__lowerCamelCase ): if valid_coloring(graph[index] , __lowerCamelCase , __lowerCamelCase ): # Color current vertex __snake_case : List[str] = i # Validate coloring if util_color(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , index + 1 ): return True # Backtrack __snake_case : str = -1 return False def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = [-1] * len(__lowerCamelCase ) if util_color(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , 0 ): return colored_vertices return []
81
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version(">=", FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType _snake_case : Any = get_logger(__name__) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=0 ): os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __snake_case : Optional[Any] = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __snake_case : Tuple = F'{MODEL_NAME}.bin' if model_index == 0 else F'{MODEL_NAME}_{model_index}.bin' __snake_case : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) if accelerator.process_index == 0: logger.info(F'Saving model to {output_model_file}' ) torch.save(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Model saved to {output_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __snake_case : Optional[int] = ( F'{MODEL_NAME}_rank{accelerator.process_index}.bin' if model_index == 0 else F'{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin' ) __snake_case : Tuple = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Saving model to {output_model_file}' ) torch.save(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Model saved to {output_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'{MODEL_NAME}_{model_index}' ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) logger.info(F'Saving model to {ckpt_dir}' ) __snake_case : Optional[int] = {"model": state_dict} dist_cp.save_state_dict( state_dict=__lowerCamelCase , storage_writer=dist_cp.FileSystemWriter(__lowerCamelCase ) , planner=DefaultSavePlanner() , ) logger.info(F'Model saved to {ckpt_dir}' ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=0 ): accelerator.wait_for_everyone() with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(__lowerCamelCase ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( "Set the `sync_module_states` flag to `True` so that model states are synced across processes when " "initializing FSDP object" ) return __snake_case : List[Any] = F'{MODEL_NAME}.bin' if model_index == 0 else F'{MODEL_NAME}_{model_index}.bin' __snake_case : Dict = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Loading model from {input_model_file}' ) __snake_case : Dict = torch.load(__lowerCamelCase ) logger.info(F'Model loaded from {input_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __snake_case : Any = ( F'{MODEL_NAME}_rank{accelerator.process_index}.bin' if model_index == 0 else F'{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin' ) __snake_case : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Loading model from {input_model_file}' ) __snake_case : Any = torch.load(__lowerCamelCase ) logger.info(F'Model loaded from {input_model_file}' ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __snake_case : Optional[int] = ( os.path.join(__lowerCamelCase , F'{MODEL_NAME}_{model_index}' ) if F'{MODEL_NAME}' not in input_dir else input_dir ) logger.info(F'Loading model from {ckpt_dir}' ) __snake_case : Optional[int] = {"model": model.state_dict()} dist_cp.load_state_dict( state_dict=__lowerCamelCase , storage_reader=dist_cp.FileSystemReader(__lowerCamelCase ) , planner=DefaultLoadPlanner() , ) __snake_case : int = state_dict["model"] logger.info(F'Model loaded from {ckpt_dir}' ) model.load_state_dict(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=0 ): os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __snake_case : Dict = FSDP.optim_state_dict(__lowerCamelCase , __lowerCamelCase ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: __snake_case : List[str] = ( F'{OPTIMIZER_NAME}.bin' if optimizer_index == 0 else F'{OPTIMIZER_NAME}_{optimizer_index}.bin' ) __snake_case : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Saving Optimizer state to {output_optimizer_file}' ) torch.save(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Optimizer state saved in {output_optimizer_file}' ) else: __snake_case : Any = os.path.join(__lowerCamelCase , F'{OPTIMIZER_NAME}_{optimizer_index}' ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) logger.info(F'Saving Optimizer state to {ckpt_dir}' ) dist_cp.save_state_dict( state_dict={"optimizer": optim_state} , storage_writer=dist_cp.FileSystemWriter(__lowerCamelCase ) , planner=DefaultSavePlanner() , ) logger.info(F'Optimizer state saved in {ckpt_dir}' ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=0 ): accelerator.wait_for_everyone() with FSDP.state_dict_type( __lowerCamelCase , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __snake_case : Union[str, Any] = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: __snake_case : Dict = ( F'{OPTIMIZER_NAME}.bin' if optimizer_index == 0 else F'{OPTIMIZER_NAME}_{optimizer_index}.bin' ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , __lowerCamelCase ) logger.info(F'Loading Optimizer state from {input_optimizer_file}' ) __snake_case : Dict = torch.load(__lowerCamelCase ) logger.info(F'Optimizer state loaded from {input_optimizer_file}' ) else: __snake_case : Optional[int] = ( os.path.join(__lowerCamelCase , F'{OPTIMIZER_NAME}_{optimizer_index}' ) if F'{OPTIMIZER_NAME}' not in input_dir else input_dir ) logger.info(F'Loading Optimizer from {ckpt_dir}' ) __snake_case : str = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key="optimizer" , storage_reader=dist_cp.FileSystemReader(__lowerCamelCase ) , ) __snake_case : List[Any] = optim_state["optimizer"] logger.info(F'Optimizer loaded from {ckpt_dir}' ) __snake_case : Tuple = FSDP.optim_state_dict_to_load(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) optimizer.load_state_dict(__lowerCamelCase )
81
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Any = CustomTokenizer pass
81
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
1
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case : Dict = logging.get_logger(__name__) _snake_case : Union[str, Any] = { "vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_config_file": "tokenizer_config.json", } _snake_case : Union[str, Any] = { "vocab_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json" }, "merges_file": { "facebook/blenderbot_small-90M": "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt" }, "tokenizer_config_file": { "facebook/blenderbot_small-90M": ( "https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json" ) }, } _snake_case : int = {"facebook/blenderbot_small-90M": 512} def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = set() __snake_case : str = word[0] for char in word[1:]: pairs.add((prev_char, char) ) __snake_case : Optional[int] = char __snake_case : Dict = set(__lowerCamelCase ) return pairs class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Tuple = VOCAB_FILES_NAMES __UpperCAmelCase : int = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self : Optional[int] , lowerCamelCase : Optional[int] , lowerCamelCase : List[Any] , lowerCamelCase : Dict="__start__" , lowerCamelCase : int="__end__" , lowerCamelCase : Optional[Any]="__unk__" , lowerCamelCase : Union[str, Any]="__null__" , **lowerCamelCase : Dict , ) -> Tuple: super().__init__(unk_token=lowerCamelCase , bos_token=lowerCamelCase , eos_token=lowerCamelCase , pad_token=lowerCamelCase , **lowerCamelCase ) with open(lowerCamelCase , encoding="utf-8" ) as vocab_handle: __snake_case : Optional[int] = json.load(lowerCamelCase ) __snake_case : int = {v: k for k, v in self.encoder.items()} with open(lowerCamelCase , encoding="utf-8" ) as merges_handle: __snake_case : Union[str, Any] = merges_handle.read().split("\n" )[1:-1] __snake_case : str = [tuple(merge.split() ) for merge in merges] __snake_case : Optional[int] = dict(zip(lowerCamelCase , range(len(lowerCamelCase ) ) ) ) __snake_case : int = {} @property def __snake_case ( self : List[str] ) -> int: return len(self.encoder ) def __snake_case ( self : int ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def __snake_case ( self : str , lowerCamelCase : str ) -> str: if token in self.cache: return self.cache[token] __snake_case : str = re.sub("([.,!?()])" , R" \1" , lowerCamelCase ) __snake_case : Union[str, Any] = re.sub("(')" , R" \1 " , lowerCamelCase ) __snake_case : Optional[int] = re.sub(R"\s{2,}" , " " , lowerCamelCase ) if "\n" in token: __snake_case : Tuple = token.replace("\n" , " __newln__" ) __snake_case : Optional[Any] = token.split(" " ) __snake_case : Dict = [] for token in tokens: if not len(lowerCamelCase ): continue __snake_case : List[Any] = token.lower() __snake_case : Optional[Any] = tuple(lowerCamelCase ) __snake_case : str = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) __snake_case : Optional[Any] = get_pairs(lowerCamelCase ) if not pairs: words.append(lowerCamelCase ) continue while True: __snake_case : Optional[Any] = min(lowerCamelCase , key=lambda lowerCamelCase : self.bpe_ranks.get(lowerCamelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break __snake_case , __snake_case : Dict = bigram __snake_case : int = [] __snake_case : Optional[int] = 0 while i < len(lowerCamelCase ): try: __snake_case : Optional[int] = word.index(lowerCamelCase , lowerCamelCase ) new_word.extend(word[i:j] ) __snake_case : Optional[int] = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(lowerCamelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 __snake_case : List[Any] = tuple(lowerCamelCase ) __snake_case : Union[str, Any] = new_word if len(lowerCamelCase ) == 1: break else: __snake_case : Tuple = get_pairs(lowerCamelCase ) __snake_case : List[Any] = "@@ ".join(lowerCamelCase ) __snake_case : List[str] = word[:-4] __snake_case : List[Any] = word words.append(lowerCamelCase ) return " ".join(lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : str ) -> List[str]: __snake_case : int = [] __snake_case : Optional[int] = re.findall(R"\S+\n?" , lowerCamelCase ) for token in words: split_tokens.extend(list(self.bpe(lowerCamelCase ).split(" " ) ) ) return split_tokens def __snake_case ( self : Optional[Any] , lowerCamelCase : str ) -> int: __snake_case : str = token.lower() return self.encoder.get(lowerCamelCase , self.encoder.get(self.unk_token ) ) def __snake_case ( self : List[Any] , lowerCamelCase : int ) -> str: return self.decoder.get(lowerCamelCase , self.unk_token ) def __snake_case ( self : Dict , lowerCamelCase : List[str] ) -> str: __snake_case : List[str] = " ".join(lowerCamelCase ).replace("@@ " , "" ).strip() return out_string def __snake_case ( self : Tuple , lowerCamelCase : str , lowerCamelCase : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCamelCase ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __snake_case : int = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) __snake_case : Tuple = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCamelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase , ensure_ascii=lowerCamelCase ) + "\n" ) __snake_case : Tuple = 0 with open(lowerCamelCase , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' " Please check that the tokenizer is not corrupted!" ) __snake_case : Dict = token_index writer.write(" ".join(lowerCamelCase ) + "\n" ) index += 1 return vocab_file, merge_file
81
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
1
import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Any ) -> List[Any]: __snake_case : Dict = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __snake_case : Optional[Any] = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(lowerCamelCase ) __snake_case : List[str] = -1 __snake_case : Union[str, Any] = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowerCamelCase ) __snake_case : Union[str, Any] = model.generate(lowerCamelCase , max_new_tokens=10 , do_sample=lowerCamelCase ) __snake_case : List[str] = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: __snake_case : Union[str, Any] = TextStreamer(lowerCamelCase ) model.generate(lowerCamelCase , max_new_tokens=10 , do_sample=lowerCamelCase , streamer=lowerCamelCase ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __snake_case : Union[str, Any] = cs.out[:-1] self.assertEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Tuple ) -> str: __snake_case : List[Any] = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __snake_case : Optional[int] = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(lowerCamelCase ) __snake_case : Optional[int] = -1 __snake_case : int = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowerCamelCase ) __snake_case : Tuple = model.generate(lowerCamelCase , max_new_tokens=10 , do_sample=lowerCamelCase ) __snake_case : List[Any] = tokenizer.decode(greedy_ids[0] ) __snake_case : Tuple = TextIteratorStreamer(lowerCamelCase ) __snake_case : List[Any] = {"input_ids": input_ids, "max_new_tokens": 10, "do_sample": False, "streamer": streamer} __snake_case : Union[str, Any] = Thread(target=model.generate , kwargs=lowerCamelCase ) thread.start() __snake_case : Optional[Any] = "" for new_text in streamer: streamer_text += new_text self.assertEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Any = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __snake_case : Dict = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(lowerCamelCase ) __snake_case : Any = -1 __snake_case : Optional[int] = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowerCamelCase ) __snake_case : Dict = model.generate(lowerCamelCase , max_new_tokens=10 , do_sample=lowerCamelCase ) __snake_case : Optional[Any] = greedy_ids[:, input_ids.shape[1] :] __snake_case : List[str] = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: __snake_case : Any = TextStreamer(lowerCamelCase , skip_prompt=lowerCamelCase ) model.generate(lowerCamelCase , max_new_tokens=10 , do_sample=lowerCamelCase , streamer=lowerCamelCase ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer __snake_case : str = cs.out[:-1] self.assertEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> List[Any]: # Tests that we can pass `decode_kwargs` to the streamer to control how the tokens are decoded. Must be tested # with actual models -- the dummy models' tokenizers are not aligned with their models, and # `skip_special_tokens=True` has no effect on them __snake_case : Optional[Any] = AutoTokenizer.from_pretrained("distilgpt2" ) __snake_case : Tuple = AutoModelForCausalLM.from_pretrained("distilgpt2" ).to(lowerCamelCase ) __snake_case : List[Any] = -1 __snake_case : Union[str, Any] = torch.ones((1, 5) , device=lowerCamelCase ).long() * model.config.bos_token_id with CaptureStdout() as cs: __snake_case : Tuple = TextStreamer(lowerCamelCase , skip_special_tokens=lowerCamelCase ) model.generate(lowerCamelCase , max_new_tokens=1 , do_sample=lowerCamelCase , streamer=lowerCamelCase ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token __snake_case : Any = cs.out[:-1] # Remove the final "\n" __snake_case : Optional[Any] = tokenizer(lowerCamelCase , return_tensors="pt" ) self.assertEqual(streamer_text_tokenized.input_ids.shape , (1, 1) ) def __snake_case ( self : Any ) -> int: __snake_case : str = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) __snake_case : Union[str, Any] = AutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-random-gpt2" ).to(lowerCamelCase ) __snake_case : Optional[Any] = -1 __snake_case : Optional[Any] = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowerCamelCase ) __snake_case : List[Any] = TextIteratorStreamer(lowerCamelCase , timeout=0.0_01 ) __snake_case : Any = {"input_ids": input_ids, "max_new_tokens": 10, "do_sample": False, "streamer": streamer} __snake_case : Tuple = Thread(target=model.generate , kwargs=lowerCamelCase ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(lowerCamelCase ): __snake_case : Any = "" for new_text in streamer: streamer_text += new_text
81
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
1
from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = 1 / sqrt(2 ) ): __snake_case : List[Any] = tau * frequency / samplerate __snake_case : Union[str, Any] = sin(__lowerCamelCase ) __snake_case : int = cos(__lowerCamelCase ) __snake_case : Union[str, Any] = _sin / (2 * q_factor) __snake_case : str = (1 - _cos) / 2 __snake_case : List[Any] = 1 - _cos __snake_case : Any = 1 + alpha __snake_case : List[str] = -2 * _cos __snake_case : List[str] = 1 - alpha __snake_case : str = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = 1 / sqrt(2 ) ): __snake_case : Dict = tau * frequency / samplerate __snake_case : List[str] = sin(__lowerCamelCase ) __snake_case : Dict = cos(__lowerCamelCase ) __snake_case : int = _sin / (2 * q_factor) __snake_case : Tuple = (1 + _cos) / 2 __snake_case : Tuple = -1 - _cos __snake_case : List[Any] = 1 + alpha __snake_case : Optional[Any] = -2 * _cos __snake_case : Optional[Any] = 1 - alpha __snake_case : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = 1 / sqrt(2 ) ): __snake_case : Union[str, Any] = tau * frequency / samplerate __snake_case : Optional[Any] = sin(__lowerCamelCase ) __snake_case : Optional[Any] = cos(__lowerCamelCase ) __snake_case : Tuple = _sin / (2 * q_factor) __snake_case : Any = _sin / 2 __snake_case : Tuple = 0 __snake_case : Optional[int] = -ba __snake_case : List[str] = 1 + alpha __snake_case : Dict = -2 * _cos __snake_case : int = 1 - alpha __snake_case : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = 1 / sqrt(2 ) ): __snake_case : Tuple = tau * frequency / samplerate __snake_case : Union[str, Any] = sin(__lowerCamelCase ) __snake_case : str = cos(__lowerCamelCase ) __snake_case : Any = _sin / (2 * q_factor) __snake_case : Optional[Any] = 1 - alpha __snake_case : Optional[int] = -2 * _cos __snake_case : Optional[Any] = 1 + alpha __snake_case : List[Any] = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = 1 / sqrt(2 ) , ): __snake_case : Tuple = tau * frequency / samplerate __snake_case : List[str] = sin(__lowerCamelCase ) __snake_case : List[Any] = cos(__lowerCamelCase ) __snake_case : Optional[int] = _sin / (2 * q_factor) __snake_case : Optional[int] = 1_0 ** (gain_db / 4_0) __snake_case : Optional[int] = 1 + alpha * big_a __snake_case : Dict = -2 * _cos __snake_case : Optional[Any] = 1 - alpha * big_a __snake_case : Optional[int] = 1 + alpha / big_a __snake_case : Union[str, Any] = -2 * _cos __snake_case : Optional[Any] = 1 - alpha / big_a __snake_case : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = 1 / sqrt(2 ) , ): __snake_case : Tuple = tau * frequency / samplerate __snake_case : int = sin(__lowerCamelCase ) __snake_case : Union[str, Any] = cos(__lowerCamelCase ) __snake_case : str = _sin / (2 * q_factor) __snake_case : Union[str, Any] = 1_0 ** (gain_db / 4_0) __snake_case : Dict = (big_a + 1) - (big_a - 1) * _cos __snake_case : Tuple = (big_a + 1) + (big_a - 1) * _cos __snake_case : int = (big_a - 1) - (big_a + 1) * _cos __snake_case : Tuple = (big_a - 1) + (big_a + 1) * _cos __snake_case : Any = 2 * sqrt(__lowerCamelCase ) * alpha __snake_case : Any = big_a * (pmc + aaa) __snake_case : Any = 2 * big_a * mpc __snake_case : Any = big_a * (pmc - aaa) __snake_case : List[Any] = ppmc + aaa __snake_case : List[str] = -2 * pmpc __snake_case : str = ppmc - aaa __snake_case : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = 1 / sqrt(2 ) , ): __snake_case : List[Any] = tau * frequency / samplerate __snake_case : Dict = sin(__lowerCamelCase ) __snake_case : List[Any] = cos(__lowerCamelCase ) __snake_case : int = _sin / (2 * q_factor) __snake_case : str = 1_0 ** (gain_db / 4_0) __snake_case : Optional[int] = (big_a + 1) - (big_a - 1) * _cos __snake_case : List[str] = (big_a + 1) + (big_a - 1) * _cos __snake_case : Tuple = (big_a - 1) - (big_a + 1) * _cos __snake_case : Any = (big_a - 1) + (big_a + 1) * _cos __snake_case : str = 2 * sqrt(__lowerCamelCase ) * alpha __snake_case : Any = big_a * (ppmc + aaa) __snake_case : Tuple = -2 * big_a * pmpc __snake_case : Tuple = big_a * (ppmc - aaa) __snake_case : Dict = pmc + aaa __snake_case : Tuple = 2 * mpc __snake_case : List[Any] = pmc - aaa __snake_case : Optional[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
81
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
1
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zstandard def lowerCAmelCase_ ( __lowerCamelCase ): assert "mock" in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ): assert "mock" not in _fsspec_registry assert "bz2" in _fsspec_registry def lowerCAmelCase_ ( ): __snake_case : Any = "mock-s3-bucket" __snake_case : str = F's3://{mock_bucket}' __snake_case : Optional[Any] = extract_path_from_uri(__lowerCamelCase ) assert dataset_path.startswith("s3://" ) is False __snake_case : Tuple = "./local/path" __snake_case : Tuple = extract_path_from_uri(__lowerCamelCase ) assert dataset_path == new_dataset_path def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = is_remote_filesystem(__lowerCamelCase ) assert is_remote is True __snake_case : Tuple = fsspec.filesystem("file" ) __snake_case : Optional[int] = is_remote_filesystem(__lowerCamelCase ) assert is_remote is False @pytest.mark.parametrize("compression_fs_class" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_file, "bz2": bza_file, "lz4": lza_file} __snake_case : Union[str, Any] = input_paths[compression_fs_class.protocol] if input_path is None: __snake_case : Tuple = F'for \'{compression_fs_class.protocol}\' compression protocol, ' if compression_fs_class.protocol == "lz4": reason += require_lza.kwargs["reason"] elif compression_fs_class.protocol == "zstd": reason += require_zstandard.kwargs["reason"] pytest.skip(__lowerCamelCase ) __snake_case : int = fsspec.filesystem(compression_fs_class.protocol , fo=__lowerCamelCase ) assert isinstance(__lowerCamelCase , __lowerCamelCase ) __snake_case : Optional[Any] = os.path.basename(__lowerCamelCase ) __snake_case : Dict = expected_filename[: expected_filename.rindex("." )] assert fs.glob("*" ) == [expected_filename] with fs.open(__lowerCamelCase , "r" , encoding="utf-8" ) as f, open(__lowerCamelCase , encoding="utf-8" ) as expected_file: assert f.read() == expected_file.read() @pytest.mark.parametrize("protocol" , ["zip", "gzip"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Union[str, Any] = {"zip": zip_jsonl_path, "gzip": jsonl_gz_path} __snake_case : int = compressed_file_paths[protocol] __snake_case : str = "dataset.jsonl" __snake_case : List[str] = F'{protocol}://{member_file_path}::{compressed_file_path}' __snake_case , *__snake_case : Dict = fsspec.get_fs_token_paths(__lowerCamelCase ) assert fs.isfile(__lowerCamelCase ) assert not fs.isfile("non_existing_" + member_file_path ) @pytest.mark.integration def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[Any] = hf_api.dataset_info(__lowerCamelCase , token=__lowerCamelCase ) __snake_case : List[str] = HfFileSystem(repo_info=__lowerCamelCase , token=__lowerCamelCase ) assert sorted(hffs.glob("*" ) ) == [".gitattributes", "data"] assert hffs.isdir("data" ) assert hffs.isfile(".gitattributes" ) and hffs.isfile("data/text_data.txt" ) with open(__lowerCamelCase ) as f: assert hffs.open("data/text_data.txt" , "r" ).read() == f.read() def lowerCAmelCase_ ( ): __snake_case : Optional[int] = "bz2" # Import module import datasets.filesystems # Overwrite protocol and reload register_implementation(__lowerCamelCase , __lowerCamelCase , clobber=__lowerCamelCase ) with pytest.warns(__lowerCamelCase ) as warning_info: importlib.reload(datasets.filesystems ) assert len(__lowerCamelCase ) == 1 assert ( str(warning_info[0].message ) == F'A filesystem protocol was already set for {protocol} and will be overwritten.' )
81
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
1
import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class a (unittest.TestCase ): """simple docstring""" @slow def __snake_case ( self : Optional[Any] ) -> List[Any]: __snake_case : List[str] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small" ) __snake_case : Tuple = AutoTokenizer.from_pretrained("google/mt5-small" ) __snake_case : Optional[Any] = tokenizer("Hello there" , return_tensors="np" ).input_ids __snake_case : List[Any] = tokenizer("Hi I am" , return_tensors="np" ).input_ids __snake_case : Union[str, Any] = shift_tokens_right(lowerCamelCase , model.config.pad_token_id , model.config.decoder_start_token_id ) __snake_case : List[str] = model(lowerCamelCase , decoder_input_ids=lowerCamelCase ).logits __snake_case : List[str] = optax.softmax_cross_entropy(lowerCamelCase , onehot(lowerCamelCase , logits.shape[-1] ) ).mean() __snake_case : Optional[int] = -(labels.shape[-1] * loss.item()) __snake_case : List[Any] = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
81
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
1
import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() _snake_case : Union[str, Any] = logging.get_logger(__name__) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[Any] = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: __snake_case : Union[str, Any] = 1_2_8 elif "12-12" in model_name: __snake_case : str = 1_2 __snake_case : Optional[int] = 1_2 elif "14-14" in model_name: __snake_case : List[Any] = 1_4 __snake_case : Dict = 1_4 elif "16-16" in model_name: __snake_case : Optional[int] = 1_6 __snake_case : Optional[int] = 1_6 else: raise ValueError("Model not supported" ) __snake_case : Optional[int] = "huggingface/label-files" if "speech-commands" in model_name: __snake_case : Tuple = 3_5 __snake_case : Optional[int] = "speech-commands-v2-id2label.json" else: __snake_case : List[Any] = 5_2_7 __snake_case : Tuple = "audioset-id2label.json" __snake_case : Optional[int] = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type="dataset" ) , "r" ) ) __snake_case : Optional[int] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} __snake_case : str = idalabel __snake_case : Tuple = {v: k for k, v in idalabel.items()} return config def lowerCAmelCase_ ( __lowerCamelCase ): if "module.v" in name: __snake_case : List[Any] = name.replace("module.v" , "audio_spectrogram_transformer" ) if "cls_token" in name: __snake_case : Any = name.replace("cls_token" , "embeddings.cls_token" ) if "dist_token" in name: __snake_case : List[Any] = name.replace("dist_token" , "embeddings.distillation_token" ) if "pos_embed" in name: __snake_case : Tuple = name.replace("pos_embed" , "embeddings.position_embeddings" ) if "patch_embed.proj" in name: __snake_case : Dict = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) # transformer blocks if "blocks" in name: __snake_case : Union[str, Any] = name.replace("blocks" , "encoder.layer" ) if "attn.proj" in name: __snake_case : List[Any] = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: __snake_case : Optional[int] = name.replace("attn" , "attention.self" ) if "norm1" in name: __snake_case : Any = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: __snake_case : int = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: __snake_case : List[str] = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: __snake_case : Union[str, Any] = name.replace("mlp.fc2" , "output.dense" ) # final layernorm if "audio_spectrogram_transformer.norm" in name: __snake_case : Union[str, Any] = name.replace("audio_spectrogram_transformer.norm" , "audio_spectrogram_transformer.layernorm" ) # classifier head if "module.mlp_head.0" in name: __snake_case : List[str] = name.replace("module.mlp_head.0" , "classifier.layernorm" ) if "module.mlp_head.1" in name: __snake_case : str = name.replace("module.mlp_head.1" , "classifier.dense" ) return name def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): for key in orig_state_dict.copy().keys(): __snake_case : List[str] = orig_state_dict.pop(__lowerCamelCase ) if "qkv" in key: __snake_case : Optional[int] = key.split("." ) __snake_case : int = int(key_split[3] ) __snake_case : str = config.hidden_size if "weight" in key: __snake_case : int = val[:dim, :] __snake_case : Optional[Any] = val[dim : dim * 2, :] __snake_case : int = val[-dim:, :] else: __snake_case : Union[str, Any] = val[:dim] __snake_case : int = val[dim : dim * 2] __snake_case : Any = val[-dim:] else: __snake_case : Optional[int] = val return orig_state_dict def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = [ "module.v.head.weight", "module.v.head.bias", "module.v.head_dist.weight", "module.v.head_dist.bias", ] for k in ignore_keys: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) @torch.no_grad() def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=False ): __snake_case : Optional[Any] = get_audio_spectrogram_transformer_config(__lowerCamelCase ) __snake_case : List[str] = { "ast-finetuned-audioset-10-10-0.4593": ( "https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.450": ( "https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.448": ( "https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.448-v2": ( "https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1" ), "ast-finetuned-audioset-12-12-0.447": ( "https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1" ), "ast-finetuned-audioset-14-14-0.443": ( "https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1" ), "ast-finetuned-audioset-16-16-0.442": ( "https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1" ), "ast-finetuned-speech-commands-v2": ( "https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1" ), } # load original state_dict __snake_case : List[Any] = model_name_to_url[model_name] __snake_case : List[str] = torch.hub.load_state_dict_from_url(__lowerCamelCase , map_location="cpu" ) # remove some keys remove_keys(__lowerCamelCase ) # rename some keys __snake_case : Tuple = convert_state_dict(__lowerCamelCase , __lowerCamelCase ) # load 🤗 model __snake_case : Union[str, Any] = ASTForAudioClassification(__lowerCamelCase ) model.eval() model.load_state_dict(__lowerCamelCase ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 __snake_case : Union[str, Any] = -4.2_6_7_7_3_9_3 if "speech-commands" not in model_name else -6.8_4_5_9_7_8 __snake_case : Tuple = 4.5_6_8_9_9_7_4 if "speech-commands" not in model_name else 5.5_6_5_4_5_2_6 __snake_case : str = 1_0_2_4 if "speech-commands" not in model_name else 1_2_8 __snake_case : Tuple = ASTFeatureExtractor(mean=__lowerCamelCase , std=__lowerCamelCase , max_length=__lowerCamelCase ) if "speech-commands" in model_name: __snake_case : Dict = load_dataset("speech_commands" , "v0.02" , split="validation" ) __snake_case : Optional[int] = dataset[0]["audio"]["array"] else: __snake_case : int = hf_hub_download( repo_id="nielsr/audio-spectogram-transformer-checkpoint" , filename="sample_audio.flac" , repo_type="dataset" , ) __snake_case , __snake_case : Optional[Any] = torchaudio.load(__lowerCamelCase ) __snake_case : Optional[Any] = waveform.squeeze().numpy() __snake_case : List[str] = feature_extractor(__lowerCamelCase , sampling_rate=1_6_0_0_0 , return_tensors="pt" ) # forward pass __snake_case : List[Any] = model(**__lowerCamelCase ) __snake_case : Tuple = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": __snake_case : List[str] = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": __snake_case : List[str] = torch.tensor([-1.1_9_8_6, -7.0_9_0_3, -8.2_7_1_8] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": __snake_case : Optional[int] = torch.tensor([-2.6_1_2_8, -8.0_0_8_0, -9.4_3_4_4] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": __snake_case : Optional[Any] = torch.tensor([-1.5_0_8_0, -7.4_5_3_4, -8.8_9_1_7] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": __snake_case : Optional[Any] = torch.tensor([-0.5_0_5_0, -6.5_8_3_3, -8.0_8_4_3] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": __snake_case : Optional[int] = torch.tensor([-0.3_8_2_6, -7.0_3_3_6, -8.2_4_1_3] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": __snake_case : int = torch.tensor([-1.2_1_1_3, -6.9_1_0_1, -8.3_4_7_0] ) elif model_name == "ast-finetuned-speech-commands-v2": __snake_case : List[str] = torch.tensor([6.1_5_8_9, -8.0_5_6_6, -8.7_9_8_4] ) else: raise ValueError("Unknown model name" ) if not torch.allclose(logits[0, :3] , __lowerCamelCase , atol=1e-4 ): raise ValueError("Logits don't match" ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__lowerCamelCase ) print(F'Saving feature extractor to {pytorch_dump_folder_path}' ) feature_extractor.save_pretrained(__lowerCamelCase ) if push_to_hub: print("Pushing model and feature extractor to the hub..." ) model.push_to_hub(F'MIT/{model_name}' ) feature_extractor.push_to_hub(F'MIT/{model_name}' ) if __name__ == "__main__": _snake_case : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="ast-finetuned-audioset-10-10-0.4593", type=str, help="Name of the Audio Spectrogram Transformer model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) _snake_case : Optional[int] = parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
81
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
1
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging _snake_case : List[str] = logging.get_logger(__name__) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = R"\w+[.]\d+" __snake_case : Dict = re.findall(__lowerCamelCase , __lowerCamelCase ) for pat in pats: __snake_case : Optional[Any] = key.replace(__lowerCamelCase , "_".join(pat.split("." ) ) ) return key def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Union[str, Any] = pt_tuple_key[:-1] + ("scale",) if ( any("norm" in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): __snake_case : List[str] = pt_tuple_key[:-1] + ("scale",) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: __snake_case : Optional[int] = pt_tuple_key[:-1] + ("scale",) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: __snake_case : Dict = pt_tuple_key[:-1] + ("embedding",) return renamed_pt_tuple_key, pt_tensor # conv layer __snake_case : Any = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: __snake_case : List[str] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __snake_case : Tuple = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight": __snake_case : Dict = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __snake_case : Tuple = pt_tuple_key[:-1] + ("weight",) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __snake_case : Optional[Any] = pt_tuple_key[:-1] + ("bias",) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=4_2 ): # Step 1: Convert pytorch tensor to numpy __snake_case : str = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params __snake_case : str = flax_model.init_weights(PRNGKey(__lowerCamelCase ) ) __snake_case : Union[str, Any] = flatten_dict(__lowerCamelCase ) __snake_case : List[Any] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __snake_case : List[str] = rename_key(__lowerCamelCase ) __snake_case : Optional[int] = tuple(renamed_pt_key.split("." ) ) # Correctly rename weight parameters __snake_case , __snake_case : List[str] = rename_key_and_reshape_tensor(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' F'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # also add unexpected weight so that warning is thrown __snake_case : Dict = jnp.asarray(__lowerCamelCase ) return unflatten_dict(__lowerCamelCase )
81
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : int , lowerCamelCase : int ) -> List[Any]: if dst_width < 0 or dst_height < 0: raise ValueError("Destination width/height should be > 0" ) __snake_case : Optional[Any] = img __snake_case : Optional[Any] = img.shape[1] __snake_case : int = img.shape[0] __snake_case : Dict = dst_width __snake_case : Union[str, Any] = dst_height __snake_case : Optional[int] = self.src_w / self.dst_w __snake_case : List[str] = self.src_h / self.dst_h __snake_case : str = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta ) * 255 ) def __snake_case ( self : Dict ) -> Dict: for i in range(self.dst_h ): for j in range(self.dst_w ): __snake_case : Union[str, Any] = self.img[self.get_y(lowerCamelCase )][self.get_x(lowerCamelCase )] def __snake_case ( self : Optional[int] , lowerCamelCase : int ) -> int: return int(self.ratio_x * x ) def __snake_case ( self : Optional[int] , lowerCamelCase : int ) -> int: return int(self.ratio_y * y ) if __name__ == "__main__": _snake_case , _snake_case : str = 800, 600 _snake_case : List[Any] = imread("image_data/lena.jpg", 1) _snake_case : Optional[Any] = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( f'''Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}''', n.output ) waitKey(0) destroyAllWindows()
81
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
1
import os from collections.abc import Iterator def lowerCAmelCase_ ( __lowerCamelCase = "." ): for dir_path, dir_names, filenames in os.walk(__lowerCamelCase ): __snake_case : Tuple = [d for d in dir_names if d != "scripts" and d[0] not in "._"] for filename in filenames: if filename == "__init__.py": continue if os.path.splitext(__lowerCamelCase )[1] in (".py", ".ipynb"): yield os.path.join(__lowerCamelCase , __lowerCamelCase ).lstrip("./" ) def lowerCAmelCase_ ( __lowerCamelCase ): return F'{i * " "}*' if i else "\n##" def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = old_path.split(os.sep ) for i, new_part in enumerate(new_path.split(os.sep ) ): if (i + 1 > len(__lowerCamelCase ) or old_parts[i] != new_part) and new_part: print(F'{md_prefix(__lowerCamelCase )} {new_part.replace("_" , " " ).title()}' ) return new_path def lowerCAmelCase_ ( __lowerCamelCase = "." ): __snake_case : Any = "" for filepath in sorted(good_file_paths(__lowerCamelCase ) ): __snake_case , __snake_case : Any = os.path.split(__lowerCamelCase ) if filepath != old_path: __snake_case : str = print_path(__lowerCamelCase , __lowerCamelCase ) __snake_case : Union[str, Any] = (filepath.count(os.sep ) + 1) if filepath else 0 __snake_case : List[Any] = F'{filepath}/{filename}'.replace(" " , "%20" ) __snake_case : str = os.path.splitext(filename.replace("_" , " " ).title() )[0] print(F'{md_prefix(__lowerCamelCase )} [{filename}]({url})' ) if __name__ == "__main__": print_directory_md(".")
81
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
import argparse from torch import nn # transformers_old should correspond to branch `save_old_prophetnet_model_structure` here # original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively from transformers_old.modeling_prophetnet import ( ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, ) from transformers_old.modeling_xlm_prophetnet import ( XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, ) from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging _snake_case : str = logging.get_logger(__name__) logging.set_verbosity_info() def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): if "xprophetnet" in prophetnet_checkpoint_path: __snake_case : Any = XLMProphetNetForConditionalGenerationOld.from_pretrained(__lowerCamelCase ) __snake_case , __snake_case : Optional[Any] = XLMProphetNetForConditionalGeneration.from_pretrained( __lowerCamelCase , output_loading_info=__lowerCamelCase ) else: __snake_case : Optional[int] = ProphetNetForConditionalGenerationOld.from_pretrained(__lowerCamelCase ) __snake_case , __snake_case : Dict = ProphetNetForConditionalGeneration.from_pretrained( __lowerCamelCase , output_loading_info=__lowerCamelCase ) __snake_case : Union[str, Any] = ["key_proj", "value_proj", "query_proj"] __snake_case : Tuple = { "self_attn": "ngram_self_attn", "cross_attn": "encoder_attn", "cross_attn_layer_norm": "encoder_attn_layer_norm", "feed_forward_layer_norm": "final_layer_norm", "feed_forward": "", "intermediate": "fc1", "output": "fc2", "key_proj": "k_proj", "query_proj": "q_proj", "value_proj": "v_proj", "word_embeddings": "embed_tokens", "embeddings_layer_norm": "emb_layer_norm", "relative_pos_embeddings": "relative_linear", "ngram_embeddings": "ngram_input_embed", "position_embeddings": "embed_positions", } for key in loading_info["missing_keys"]: __snake_case : Optional[Any] = key.split("." ) if attributes[0] == "lm_head": __snake_case : Dict = prophet __snake_case : Tuple = prophet_old else: __snake_case : str = prophet.prophetnet __snake_case : Any = prophet_old.model __snake_case : Any = False for attribute in attributes: if attribute in mapping: __snake_case : Optional[int] = mapping[attribute] if not hasattr(__lowerCamelCase , __lowerCamelCase ) and len(__lowerCamelCase ) > 0: __snake_case : List[str] = attribute elif hasattr(__lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = attribute if attribute == "weight": assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" __snake_case : Dict = old_model.weight logger.info(F'{attribute} is initialized.' ) __snake_case : int = True break elif attribute == "bias": assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" __snake_case : int = old_model.bias logger.info(F'{attribute} is initialized' ) __snake_case : Dict = True break elif attribute in special_keys and hasattr(__lowerCamelCase , "in_proj_weight" ): __snake_case : Any = old_model.in_proj_weight.shape[0] // 3 __snake_case : Dict = getattr(__lowerCamelCase , __lowerCamelCase ) param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" if attribute == "query_proj": __snake_case : Any = nn.Parameter(old_model.in_proj_weight[:embed_dim, :] ) __snake_case : Tuple = nn.Parameter(old_model.in_proj_bias[:embed_dim] ) elif attribute == "key_proj": __snake_case : str = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :] ) __snake_case : List[Any] = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim] ) elif attribute == "value_proj": __snake_case : List[Any] = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :] ) __snake_case : Optional[Any] = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :] ) __snake_case : int = True break elif attribute == "position_embeddings": assert ( model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] ), "Hidden size has to match" assert model.position_embeddings.weight.shape[0] == 5_1_2, "We want 512 position_embeddings." __snake_case : Tuple = nn.Parameter(old_model.embed_positions.weight[:5_1_2, :] ) __snake_case : List[str] = True break if attribute.isdigit(): __snake_case : List[str] = model[int(__lowerCamelCase )] __snake_case : int = old_model[int(__lowerCamelCase )] else: __snake_case : Optional[int] = getattr(__lowerCamelCase , __lowerCamelCase ) if old_attribute == "": __snake_case : Optional[Any] = old_model else: if not hasattr(__lowerCamelCase , __lowerCamelCase ): raise ValueError(F'{old_model} does not have {old_attribute}' ) __snake_case : str = getattr(__lowerCamelCase , __lowerCamelCase ) if not is_key_init: raise ValueError(F'{key} was not correctly initialized!' ) print(F'Saving model to {pytorch_dump_folder_path}' ) prophet.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _snake_case : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( "--prophetnet_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _snake_case : int = parser.parse_args() convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path)
81
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
1
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device _snake_case : Tuple = False class a (unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Any ) -> Dict: __snake_case : Optional[Any] = VersatileDiffusionTextToImagePipeline.from_pretrained("shi-labs/versatile-diffusion" ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __snake_case : Optional[int] = "A painting of a squirrel eating a burger " __snake_case : str = torch.manual_seed(0 ) __snake_case : str = pipe( prompt=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase ) __snake_case : str = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCamelCase ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __snake_case : Tuple = generator.manual_seed(0 ) __snake_case : int = pipe( prompt=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type="numpy" ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def __snake_case ( self : Dict ) -> Dict: __snake_case : Any = VersatileDiffusionTextToImagePipeline.from_pretrained( "shi-labs/versatile-diffusion" , torch_dtype=torch.floataa ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __snake_case : Optional[int] = "A painting of a squirrel eating a burger " __snake_case : Tuple = torch.manual_seed(0 ) __snake_case : int = pipe( prompt=lowerCamelCase , generator=lowerCamelCase , guidance_scale=7.5 , num_inference_steps=50 , output_type="numpy" ).images __snake_case : int = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) __snake_case : List[Any] = np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
81
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _snake_case : str = { "configuration_vision_text_dual_encoder": ["VisionTextDualEncoderConfig"], "processing_vision_text_dual_encoder": ["VisionTextDualEncoderProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = ["VisionTextDualEncoderModel"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = ["FlaxVisionTextDualEncoderModel"] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["TFVisionTextDualEncoderModel"] if TYPE_CHECKING: from .configuration_vision_text_dual_encoder import VisionTextDualEncoderConfig from .processing_vision_text_dual_encoder import VisionTextDualEncoderProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_text_dual_encoder import VisionTextDualEncoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_text_dual_encoder import FlaxVisionTextDualEncoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_text_dual_encoder import TFVisionTextDualEncoderModel else: import sys _snake_case : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure)
81
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
1
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils import IMAGE_PROCESSOR_NAME, is_vision_available if is_vision_available(): from PIL import Image from transformers import CLIPImageProcessor, CLIPProcessor @require_vision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Optional[Any] ) -> str: __snake_case : Dict = tempfile.mkdtemp() # fmt: off __snake_case : int = ["l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "lo", "l</w>", "w</w>", "r</w>", "t</w>", "low</w>", "er</w>", "lowest</w>", "newer</w>", "wider", "<unk>", "<|startoftext|>", "<|endoftext|>"] # fmt: on __snake_case : List[str] = dict(zip(lowerCamelCase , range(len(lowerCamelCase ) ) ) ) __snake_case : Any = ["#version: 0.2", "l o", "lo w</w>", "e r</w>", ""] __snake_case : str = {"unk_token": "<unk>"} __snake_case : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __snake_case : Any = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCamelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(lowerCamelCase ) ) __snake_case : Optional[Any] = { "do_resize": True, "size": 20, "do_center_crop": True, "crop_size": 18, "do_normalize": True, "image_mean": [0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73], "image_std": [0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11], } __snake_case : List[Any] = os.path.join(self.tmpdirname , lowerCamelCase ) with open(self.image_processor_file , "w" , encoding="utf-8" ) as fp: json.dump(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Dict , **lowerCamelCase : str ) -> str: return CLIPTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase ) def __snake_case ( self : List[Any] , **lowerCamelCase : Optional[int] ) -> Optional[int]: return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **lowerCamelCase ) def __snake_case ( self : List[str] , **lowerCamelCase : Optional[int] ) -> Optional[int]: return CLIPImageProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> str: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Tuple ) -> Union[str, Any]: __snake_case : Any = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Optional[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[str] = self.get_tokenizer() __snake_case : str = self.get_rust_tokenizer() __snake_case : int = self.get_image_processor() __snake_case : Dict = CLIPProcessor(tokenizer=lowerCamelCase , image_processor=lowerCamelCase ) processor_slow.save_pretrained(self.tmpdirname ) __snake_case : Optional[int] = CLIPProcessor.from_pretrained(self.tmpdirname , use_fast=lowerCamelCase ) __snake_case : Dict = CLIPProcessor(tokenizer=lowerCamelCase , image_processor=lowerCamelCase ) processor_fast.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = CLIPProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor_slow.tokenizer.get_vocab() , tokenizer_slow.get_vocab() ) self.assertEqual(processor_fast.tokenizer.get_vocab() , tokenizer_fast.get_vocab() ) self.assertEqual(tokenizer_slow.get_vocab() , tokenizer_fast.get_vocab() ) self.assertIsInstance(processor_slow.tokenizer , lowerCamelCase ) self.assertIsInstance(processor_fast.tokenizer , lowerCamelCase ) self.assertEqual(processor_slow.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertEqual(processor_fast.image_processor.to_json_string() , image_processor.to_json_string() ) self.assertIsInstance(processor_slow.image_processor , lowerCamelCase ) self.assertIsInstance(processor_fast.image_processor , lowerCamelCase ) def __snake_case ( self : int ) -> Optional[Any]: __snake_case : str = CLIPProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : List[str] = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) __snake_case : str = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : List[str] = CLIPProcessor.from_pretrained( self.tmpdirname , bos_token="(BOS)" , eos_token="(EOS)" , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , lowerCamelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Dict ) -> str: __snake_case : List[Any] = self.get_image_processor() __snake_case : Any = self.get_tokenizer() __snake_case : List[str] = CLIPProcessor(tokenizer=lowerCamelCase , image_processor=lowerCamelCase ) __snake_case : Any = self.prepare_image_inputs() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : str = processor(images=lowerCamelCase , return_tensors="np" ) for key in input_image_proc.keys(): self.assertAlmostEqual(input_image_proc[key].sum() , input_processor[key].sum() , delta=1E-2 ) def __snake_case ( self : Dict ) -> Any: __snake_case : int = self.get_image_processor() __snake_case : List[Any] = self.get_tokenizer() __snake_case : List[str] = CLIPProcessor(tokenizer=lowerCamelCase , image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = "lower newer" __snake_case : Any = processor(text=lowerCamelCase ) __snake_case : Dict = tokenizer(lowerCamelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.get_image_processor() __snake_case : int = self.get_tokenizer() __snake_case : Optional[Any] = CLIPProcessor(tokenizer=lowerCamelCase , image_processor=lowerCamelCase ) __snake_case : Tuple = "lower newer" __snake_case : Dict = self.prepare_image_inputs() __snake_case : List[str] = processor(text=lowerCamelCase , images=lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , ["input_ids", "attention_mask", "pixel_values"] ) # test if it raises when no input is passed with pytest.raises(lowerCamelCase ): processor() def __snake_case ( self : Optional[int] ) -> Tuple: __snake_case : List[str] = self.get_image_processor() __snake_case : List[Any] = self.get_tokenizer() __snake_case : int = CLIPProcessor(tokenizer=lowerCamelCase , image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] __snake_case : Dict = processor.batch_decode(lowerCamelCase ) __snake_case : Dict = tokenizer.batch_decode(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Tuple ) -> Any: __snake_case : Any = self.get_image_processor() __snake_case : List[str] = self.get_tokenizer() __snake_case : Tuple = CLIPProcessor(tokenizer=lowerCamelCase , image_processor=lowerCamelCase ) __snake_case : Optional[Any] = "lower newer" __snake_case : List[Any] = self.prepare_image_inputs() __snake_case : Any = processor(text=lowerCamelCase , images=lowerCamelCase ) self.assertListEqual(list(inputs.keys() ) , processor.model_input_names )
81
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
1
import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 _snake_case : List[str] = sys.version_info >= (3, 10) def lowerCAmelCase_ ( __lowerCamelCase=None , __lowerCamelCase=None ): return field(default_factory=lambda: default , metadata=__lowerCamelCase ) @dataclass class a : """simple docstring""" __UpperCAmelCase : int __UpperCAmelCase : float __UpperCAmelCase : str __UpperCAmelCase : bool @dataclass class a : """simple docstring""" __UpperCAmelCase : int = 42 __UpperCAmelCase : str = field(default="toto" , metadata={"help": "help message"} ) @dataclass class a : """simple docstring""" __UpperCAmelCase : bool = False __UpperCAmelCase : bool = True __UpperCAmelCase : Optional[bool] = None class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = "titi" __UpperCAmelCase : Union[str, Any] = "toto" class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Any = "titi" __UpperCAmelCase : Any = "toto" __UpperCAmelCase : List[Any] = 42 @dataclass class a : """simple docstring""" __UpperCAmelCase : BasicEnum = "toto" def __snake_case ( self : List[Any] ) -> Optional[Any]: __snake_case : Tuple = BasicEnum(self.foo ) @dataclass class a : """simple docstring""" __UpperCAmelCase : MixedTypeEnum = "toto" def __snake_case ( self : Optional[int] ) -> str: __snake_case : List[str] = MixedTypeEnum(self.foo ) @dataclass class a : """simple docstring""" __UpperCAmelCase : Optional[int] = None __UpperCAmelCase : Optional[float] = field(default=_lowerCAmelCase , metadata={"help": "help message"} ) __UpperCAmelCase : Optional[str] = None __UpperCAmelCase : Optional[List[str]] = list_field(default=[] ) __UpperCAmelCase : Optional[List[int]] = list_field(default=[] ) @dataclass class a : """simple docstring""" __UpperCAmelCase : List[int] = list_field(default=[] ) __UpperCAmelCase : List[int] = list_field(default=[1, 2, 3] ) __UpperCAmelCase : List[str] = list_field(default=["Hallo", "Bonjour", "Hello"] ) __UpperCAmelCase : List[float] = list_field(default=[0.1, 0.2, 0.3] ) @dataclass class a : """simple docstring""" __UpperCAmelCase : List[int] = field() __UpperCAmelCase : str = field() __UpperCAmelCase : BasicEnum = field() def __snake_case ( self : int ) -> List[str]: __snake_case : Any = BasicEnum(self.required_enum ) @dataclass class a : """simple docstring""" __UpperCAmelCase : int __UpperCAmelCase : "BasicEnum" = field() __UpperCAmelCase : "Optional[bool]" = None __UpperCAmelCase : "str" = field(default="toto" , metadata={"help": "help message"} ) __UpperCAmelCase : "List[str]" = list_field(default=["Hallo", "Bonjour", "Hello"] ) if is_python_no_less_than_3_10: @dataclass class a : """simple docstring""" __UpperCAmelCase : bool = False __UpperCAmelCase : bool = True __UpperCAmelCase : bool | None = None @dataclass class a : """simple docstring""" __UpperCAmelCase : int | None = None __UpperCAmelCase : float | None = field(default=_lowerCAmelCase , metadata={"help": "help message"} ) __UpperCAmelCase : str | None = None __UpperCAmelCase : list[str] | None = list_field(default=[] ) __UpperCAmelCase : list[int] | None = list_field(default=[] ) class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Dict , lowerCamelCase : argparse.ArgumentParser , lowerCamelCase : argparse.ArgumentParser ) -> Dict: self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): __snake_case : Union[str, Any] = {k: v for k, v in vars(lowerCamelCase ).items() if k != "container"} __snake_case : str = {k: v for k, v in vars(lowerCamelCase ).items() if k != "container"} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("choices" , lowerCamelCase ) and yy.get("choices" , lowerCamelCase ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["type"](lowerCamelCase ) , yy["type"](lowerCamelCase ) ) del xx["type"], yy["type"] self.assertEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Dict: __snake_case : Dict = HfArgumentParser(lowerCamelCase ) __snake_case : Any = argparse.ArgumentParser() expected.add_argument("--foo" , type=lowerCamelCase , required=lowerCamelCase ) expected.add_argument("--bar" , type=lowerCamelCase , required=lowerCamelCase ) expected.add_argument("--baz" , type=lowerCamelCase , required=lowerCamelCase ) expected.add_argument("--flag" , type=lowerCamelCase , default=lowerCamelCase , const=lowerCamelCase , nargs="?" ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) __snake_case : List[str] = ["--foo", "1", "--baz", "quux", "--bar", "0.5"] ((__snake_case) , ) : Optional[int] = parser.parse_args_into_dataclasses(lowerCamelCase , look_for_args_file=lowerCamelCase ) self.assertFalse(example.flag ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case : Tuple = HfArgumentParser(lowerCamelCase ) __snake_case : List[str] = argparse.ArgumentParser() expected.add_argument("--foo" , default=42 , type=lowerCamelCase ) expected.add_argument("--baz" , default="toto" , type=lowerCamelCase , help="help message" ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Tuple = argparse.ArgumentParser() expected.add_argument("--foo" , type=lowerCamelCase , default=lowerCamelCase , const=lowerCamelCase , nargs="?" ) expected.add_argument("--baz" , type=lowerCamelCase , default=lowerCamelCase , const=lowerCamelCase , nargs="?" ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("--no_baz" , action="store_false" , default=lowerCamelCase , dest="baz" ) expected.add_argument("--opt" , type=lowerCamelCase , default=lowerCamelCase ) __snake_case : Optional[int] = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(lowerCamelCase ) for dataclass_type in dataclass_types: __snake_case : Optional[Any] = HfArgumentParser(lowerCamelCase ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) __snake_case : str = parser.parse_args([] ) self.assertEqual(lowerCamelCase , Namespace(foo=lowerCamelCase , baz=lowerCamelCase , opt=lowerCamelCase ) ) __snake_case : List[str] = parser.parse_args(["--foo", "--no_baz"] ) self.assertEqual(lowerCamelCase , Namespace(foo=lowerCamelCase , baz=lowerCamelCase , opt=lowerCamelCase ) ) __snake_case : int = parser.parse_args(["--foo", "--baz"] ) self.assertEqual(lowerCamelCase , Namespace(foo=lowerCamelCase , baz=lowerCamelCase , opt=lowerCamelCase ) ) __snake_case : Tuple = parser.parse_args(["--foo", "True", "--baz", "True", "--opt", "True"] ) self.assertEqual(lowerCamelCase , Namespace(foo=lowerCamelCase , baz=lowerCamelCase , opt=lowerCamelCase ) ) __snake_case : List[str] = parser.parse_args(["--foo", "False", "--baz", "False", "--opt", "False"] ) self.assertEqual(lowerCamelCase , Namespace(foo=lowerCamelCase , baz=lowerCamelCase , opt=lowerCamelCase ) ) def __snake_case ( self : str ) -> Tuple: __snake_case : Optional[int] = HfArgumentParser(lowerCamelCase ) __snake_case : Tuple = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=["titi", "toto", 42] , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) __snake_case : Union[str, Any] = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) __snake_case : int = parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) __snake_case : str = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) __snake_case : Optional[int] = parser.parse_args_into_dataclasses(["--foo", "titi"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) __snake_case : List[str] = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) __snake_case : Union[str, Any] = parser.parse_args_into_dataclasses(["--foo", "42"] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def __snake_case ( self : int ) -> int: @dataclass class a : """simple docstring""" __UpperCAmelCase : Literal["titi", "toto", 42] = "toto" __snake_case : int = HfArgumentParser(lowerCamelCase ) __snake_case : Union[str, Any] = argparse.ArgumentParser() expected.add_argument( "--foo" , default="toto" , choices=("titi", "toto", 42) , type=make_choice_type_function(["titi", "toto", 42] ) , ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[int] = parser.parse_args([] ) self.assertEqual(args.foo , "toto" ) __snake_case : Tuple = parser.parse_args(["--foo", "titi"] ) self.assertEqual(args.foo , "titi" ) __snake_case : Dict = parser.parse_args(["--foo", "42"] ) self.assertEqual(args.foo , 42 ) def __snake_case ( self : Dict ) -> List[Any]: __snake_case : str = HfArgumentParser(lowerCamelCase ) __snake_case : Any = argparse.ArgumentParser() expected.add_argument("--foo_int" , nargs="+" , default=[] , type=lowerCamelCase ) expected.add_argument("--bar_int" , nargs="+" , default=[1, 2, 3] , type=lowerCamelCase ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=lowerCamelCase ) expected.add_argument("--foo_float" , nargs="+" , default=[0.1, 0.2, 0.3] , type=lowerCamelCase ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) __snake_case : Tuple = parser.parse_args([] ) self.assertEqual( lowerCamelCase , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=["Hallo", "Bonjour", "Hello"] , foo_float=[0.1, 0.2, 0.3] ) , ) __snake_case : str = parser.parse_args("--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7".split() ) self.assertEqual(lowerCamelCase , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=["a", "b", "c"] , foo_float=[0.1, 0.7] ) ) def __snake_case ( self : str ) -> List[Any]: __snake_case : List[str] = argparse.ArgumentParser() expected.add_argument("--foo" , default=lowerCamelCase , type=lowerCamelCase ) expected.add_argument("--bar" , default=lowerCamelCase , type=lowerCamelCase , help="help message" ) expected.add_argument("--baz" , default=lowerCamelCase , type=lowerCamelCase ) expected.add_argument("--ces" , nargs="+" , default=[] , type=lowerCamelCase ) expected.add_argument("--des" , nargs="+" , default=[] , type=lowerCamelCase ) __snake_case : List[str] = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(lowerCamelCase ) for dataclass_type in dataclass_types: __snake_case : str = HfArgumentParser(lowerCamelCase ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) __snake_case : List[Any] = parser.parse_args([] ) self.assertEqual(lowerCamelCase , Namespace(foo=lowerCamelCase , bar=lowerCamelCase , baz=lowerCamelCase , ces=[] , des=[] ) ) __snake_case : List[str] = parser.parse_args("--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3".split() ) self.assertEqual(lowerCamelCase , Namespace(foo=12 , bar=3.14 , baz="42" , ces=["a", "b", "c"] , des=[1, 2, 3] ) ) def __snake_case ( self : List[str] ) -> Optional[Any]: __snake_case : int = HfArgumentParser(lowerCamelCase ) __snake_case : str = argparse.ArgumentParser() expected.add_argument("--required_list" , nargs="+" , type=lowerCamelCase , required=lowerCamelCase ) expected.add_argument("--required_str" , type=lowerCamelCase , required=lowerCamelCase ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=lowerCamelCase , ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Dict ) -> Any: __snake_case : Dict = HfArgumentParser(lowerCamelCase ) __snake_case : Tuple = argparse.ArgumentParser() expected.add_argument("--foo" , type=lowerCamelCase , required=lowerCamelCase ) expected.add_argument( "--required_enum" , type=make_choice_type_function(["titi", "toto"] ) , choices=["titi", "toto"] , required=lowerCamelCase , ) expected.add_argument("--opt" , type=lowerCamelCase , default=lowerCamelCase ) expected.add_argument("--baz" , default="toto" , type=lowerCamelCase , help="help message" ) expected.add_argument("--foo_str" , nargs="+" , default=["Hallo", "Bonjour", "Hello"] , type=lowerCamelCase ) self.argparsersEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Optional[Any]: __snake_case : int = HfArgumentParser(lowerCamelCase ) __snake_case : Optional[Any] = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } __snake_case : List[Any] = parser.parse_dict(lowerCamelCase )[0] __snake_case : int = BasicExample(**lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : int = HfArgumentParser(lowerCamelCase ) __snake_case : Any = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, "extra": 42, } self.assertRaises(lowerCamelCase , parser.parse_dict , lowerCamelCase , allow_extra_keys=lowerCamelCase ) def __snake_case ( self : int ) -> Tuple: __snake_case : Tuple = HfArgumentParser(lowerCamelCase ) __snake_case : int = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: __snake_case : Dict = os.path.join(lowerCamelCase , "temp_json" ) os.mkdir(lowerCamelCase ) with open(temp_local_path + ".json" , "w+" ) as f: json.dump(lowerCamelCase , lowerCamelCase ) __snake_case : Any = parser.parse_yaml_file(Path(temp_local_path + ".json" ) )[0] __snake_case : List[Any] = BasicExample(**lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Tuple ) -> List[Any]: __snake_case : str = HfArgumentParser(lowerCamelCase ) __snake_case : int = { "foo": 12, "bar": 3.14, "baz": "42", "flag": True, } with tempfile.TemporaryDirectory() as tmp_dir: __snake_case : Tuple = os.path.join(lowerCamelCase , "temp_yaml" ) os.mkdir(lowerCamelCase ) with open(temp_local_path + ".yaml" , "w+" ) as f: yaml.dump(lowerCamelCase , lowerCamelCase ) __snake_case : List[Any] = parser.parse_yaml_file(Path(temp_local_path + ".yaml" ) )[0] __snake_case : List[Any] = BasicExample(**lowerCamelCase ) self.assertEqual(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Tuple ) -> List[Any]: __snake_case : List[str] = HfArgumentParser(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase )
81
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
1
import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): return params[F'{prefix}/{prefix}/relpos_bias/rel_embedding'][:, i, :] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase="attention" ): __snake_case : Tuple = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/key/kernel'][:, i, :, :] ) __snake_case : Dict = k_tmp.reshape(k_tmp.shape[0] , k_tmp.shape[1] * k_tmp.shape[2] ) __snake_case : str = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/out/kernel'][:, i, :, :] ) __snake_case : List[str] = o_tmp.reshape(o_tmp.shape[0] * o_tmp.shape[1] , o_tmp.shape[2] ) __snake_case : Any = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/query/kernel'][:, i, :, :] ) __snake_case : Optional[int] = q_tmp.reshape(q_tmp.shape[0] , q_tmp.shape[1] * q_tmp.shape[2] ) __snake_case : List[str] = np.ascontiguousarray(params[F'{prefix}/{prefix}/{layer_name}/value/kernel'][:, i, :, :] ) __snake_case : List[Any] = v_tmp.reshape(v_tmp.shape[0] , v_tmp.shape[1] * v_tmp.shape[2] ) return k, o, q, v def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=False ): if split_mlp_wi: __snake_case : Any = params[F'{prefix}/{prefix}/mlp/wi_0/kernel'][:, i, :] __snake_case : List[Any] = params[F'{prefix}/{prefix}/mlp/wi_1/kernel'][:, i, :] __snake_case : Optional[Any] = (wi_a, wi_a) else: __snake_case : Optional[int] = params[F'{prefix}/{prefix}/mlp/wi/kernel'][:, i, :] __snake_case : List[Any] = params[F'{prefix}/{prefix}/mlp/wo/kernel'][:, i, :] return wi, wo def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): return params[F'{prefix}/{prefix}/{layer_name}/scale'][:, i] def lowerCAmelCase_ ( __lowerCamelCase , *, __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = False ): __snake_case : Union[str, Any] = traverse_util.flatten_dict(variables["target"] ) __snake_case : str = {"/".join(__lowerCamelCase ): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi __snake_case : Any = "encoder/encoder/mlp/wi_0/kernel" in old print("Split MLP:" , __lowerCamelCase ) __snake_case : Tuple = collections.OrderedDict() # Shared embeddings. __snake_case : Optional[int] = old["token_embedder/embedding"] # Encoder. for i in range(__lowerCamelCase ): # Block i, layer 0 (Self Attention). __snake_case : Union[str, Any] = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , "pre_attention_layer_norm" ) __snake_case , __snake_case , __snake_case , __snake_case : List[str] = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , "attention" ) __snake_case : List[str] = layer_norm __snake_case : Union[str, Any] = k.T __snake_case : List[str] = o.T __snake_case : List[str] = q.T __snake_case : Union[str, Any] = v.T # Block i, layer 1 (MLP). __snake_case : str = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , "pre_mlp_layer_norm" ) __snake_case , __snake_case : Any = tax_mlp_lookup(__lowerCamelCase , __lowerCamelCase , "encoder" , __lowerCamelCase ) __snake_case : str = layer_norm if split_mlp_wi: __snake_case : Optional[int] = wi[0].T __snake_case : List[Any] = wi[1].T else: __snake_case : Any = wi.T __snake_case : Union[str, Any] = wo.T if scalable_attention: # convert the rel_embedding of each layer __snake_case : List[Any] = tax_relpos_bias_lookup( __lowerCamelCase , __lowerCamelCase , "encoder" ).T __snake_case : int = old["encoder/encoder_norm/scale"] if not scalable_attention: __snake_case : Optional[int] = tax_relpos_bias_lookup( __lowerCamelCase , 0 , "encoder" ).T __snake_case : Union[str, Any] = tax_relpos_bias_lookup( __lowerCamelCase , 0 , "decoder" ).T if not is_encoder_only: # Decoder. for i in range(__lowerCamelCase ): # Block i, layer 0 (Self Attention). __snake_case : Tuple = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "pre_self_attention_layer_norm" ) __snake_case , __snake_case , __snake_case , __snake_case : Tuple = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "self_attention" ) __snake_case : int = layer_norm __snake_case : Tuple = k.T __snake_case : List[Any] = o.T __snake_case : str = q.T __snake_case : Optional[Any] = v.T # Block i, layer 1 (Cross Attention). __snake_case : Optional[Any] = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "pre_cross_attention_layer_norm" ) __snake_case , __snake_case , __snake_case , __snake_case : List[Any] = tax_attention_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "encoder_decoder_attention" ) __snake_case : Optional[int] = layer_norm __snake_case : int = k.T __snake_case : Optional[Any] = o.T __snake_case : Dict = q.T __snake_case : List[Any] = v.T # Block i, layer 2 (MLP). __snake_case : Dict = tax_layer_norm_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , "pre_mlp_layer_norm" ) __snake_case , __snake_case : Dict = tax_mlp_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" , __lowerCamelCase ) __snake_case : Dict = layer_norm if split_mlp_wi: __snake_case : List[Any] = wi[0].T __snake_case : int = wi[1].T else: __snake_case : Dict = wi.T __snake_case : Any = wo.T if scalable_attention: # convert the rel_embedding of each layer __snake_case : Any = tax_relpos_bias_lookup(__lowerCamelCase , __lowerCamelCase , "decoder" ).T __snake_case : Dict = old["decoder/decoder_norm/scale"] # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: __snake_case : List[Any] = old["decoder/logits_dense/kernel"].T return new def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[Any] = collections.OrderedDict([(k, torch.from_numpy(v.copy() )) for (k, v) in converted_params.items()] ) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: __snake_case : int = state_dict["shared.weight"] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: __snake_case : Optional[int] = state_dict["shared.weight"] if "lm_head.weight" not in state_dict: # For old 1.0 models. print("Using shared word embeddings as lm_head." ) __snake_case : int = state_dict["shared.weight"] return state_dict def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Union[str, Any] = checkpoints.load_tax_checkpoint(__lowerCamelCase ) __snake_case : int = convert_tax_to_pytorch( __lowerCamelCase , num_layers=config.num_layers , is_encoder_only=__lowerCamelCase , scalable_attention=__lowerCamelCase ) __snake_case : List[str] = make_state_dict(__lowerCamelCase , __lowerCamelCase ) model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = False , __lowerCamelCase = False , ): __snake_case : str = MTaConfig.from_json_file(__lowerCamelCase ) print(F'Building PyTorch model from configuration: {config}' ) # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: __snake_case : Optional[int] = UMTaEncoderModel(__lowerCamelCase ) else: __snake_case : int = UMTaForConditionalGeneration(__lowerCamelCase ) # Load weights from tf checkpoint load_tax_weights_in_ta(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Save pytorch-model print(F'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(__lowerCamelCase ) # Verify that we can load the checkpoint. model.from_pretrained(__lowerCamelCase ) print("Done" ) if __name__ == "__main__": _snake_case : Optional[Any] = argparse.ArgumentParser(description="Converts a native T5X checkpoint into a PyTorch checkpoint.") # Required parameters parser.add_argument( "--t5x_checkpoint_path", default=None, type=str, required=True, help="Path to the T5X checkpoint." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--is_encoder_only", action="store_true", help="Check if the model is encoder-decoder model", default=False ) parser.add_argument( "--scalable_attention", action="store_true", help="Whether the model uses scaled attention (umt5 model)", default=False, ) _snake_case : Optional[Any] = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only, args.scalable_attention, )
81
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
1
from math import pi, sqrt def lowerCAmelCase_ ( __lowerCamelCase ): if num <= 0: raise ValueError("math domain error" ) if num > 1_7_1.5: raise OverflowError("math range error" ) elif num - int(__lowerCamelCase ) not in (0, 0.5): raise NotImplementedError("num must be an integer or a half-integer" ) elif num == 0.5: return sqrt(__lowerCamelCase ) else: return 1.0 if num == 1 else (num - 1) * gamma(num - 1 ) def lowerCAmelCase_ ( ): assert gamma(0.5 ) == sqrt(__lowerCamelCase ) assert gamma(1 ) == 1.0 assert gamma(2 ) == 1.0 if __name__ == "__main__": from doctest import testmod testmod() _snake_case : Optional[Any] = 1.0 while num: _snake_case : Dict = float(input("Gamma of: ")) print(f'''gamma({num}) = {gamma(num)}''') print("\nEnter 0 to exit...")
81
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
1
_snake_case : dict[str, float] = { "joule": 1.0, "kilojoule": 1_000, "megajoule": 1_000_000, "gigajoule": 1_000_000_000, "wattsecond": 1.0, "watthour": 3_600, "kilowatthour": 3_600_000, "newtonmeter": 1.0, "calorie_nutr": 4_186.8, "kilocalorie_nutr": 4_186_800.00, "electronvolt": 1.6_02_17_66_34E-19, "britishthermalunit_it": 1_055.05_585, "footpound": 1.35_58_18, } def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: __snake_case : Union[str, Any] = ( F'Incorrect \'from_type\' or \'to_type\' value: {from_type!r}, {to_type!r}\n' F'Valid values are: {", ".join(__lowerCamelCase )}' ) raise ValueError(__lowerCamelCase ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
81
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
1
import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = (DDIMParallelScheduler,) __UpperCAmelCase : Tuple = (("eta", 0.0), ("num_inference_steps", 50)) def __snake_case ( self : int , **lowerCamelCase : Optional[Any] ) -> Optional[Any]: __snake_case : Any = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", "clip_sample": True, } config.update(**lowerCamelCase ) return config def __snake_case ( self : Optional[Any] , **lowerCamelCase : List[str] ) -> Any: __snake_case : Optional[int] = self.scheduler_classes[0] __snake_case : Optional[int] = self.get_scheduler_config(**lowerCamelCase ) __snake_case : Optional[int] = scheduler_class(**lowerCamelCase ) __snake_case , __snake_case : List[Any] = 10, 0.0 __snake_case : Union[str, Any] = self.dummy_model() __snake_case : Union[str, Any] = self.dummy_sample_deter scheduler.set_timesteps(lowerCamelCase ) for t in scheduler.timesteps: __snake_case : List[Any] = model(lowerCamelCase , lowerCamelCase ) __snake_case : List[str] = scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ).prev_sample return sample def __snake_case ( self : Tuple ) -> Dict: for timesteps in [100, 500, 1000]: self.check_over_configs(num_train_timesteps=lowerCamelCase ) def __snake_case ( self : Dict ) -> int: for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowerCamelCase ) __snake_case : Optional[int] = self.scheduler_classes[0] __snake_case : Optional[Any] = self.get_scheduler_config(steps_offset=1 ) __snake_case : List[str] = scheduler_class(**lowerCamelCase ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps , torch.LongTensor([801, 601, 401, 201, 1] ) ) def __snake_case ( self : List[Any] ) -> int: for beta_start, beta_end in zip([0.00_01, 0.0_01, 0.01, 0.1] , [0.0_02, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=lowerCamelCase , beta_end=lowerCamelCase ) def __snake_case ( self : List[Any] ) -> str: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowerCamelCase ) def __snake_case ( self : Tuple ) -> List[str]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> List[Any]: for clip_sample in [True, False]: self.check_over_configs(clip_sample=lowerCamelCase ) def __snake_case ( self : Tuple ) -> List[Any]: for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=lowerCamelCase ) def __snake_case ( self : Tuple ) -> Any: for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=lowerCamelCase ) def __snake_case ( self : Dict ) -> List[str]: self.check_over_configs(thresholding=lowerCamelCase ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=lowerCamelCase , prediction_type=lowerCamelCase , sample_max_value=lowerCamelCase , ) def __snake_case ( self : int ) -> Union[str, Any]: for t in [1, 10, 49]: self.check_over_forward(time_step=lowerCamelCase ) def __snake_case ( self : int ) -> List[Any]: for t, num_inference_steps in zip([1, 10, 50] , [10, 50, 500] ): self.check_over_forward(time_step=lowerCamelCase , num_inference_steps=lowerCamelCase ) def __snake_case ( self : Dict ) -> str: for t, eta in zip([1, 10, 49] , [0.0, 0.5, 1.0] ): self.check_over_forward(time_step=lowerCamelCase , eta=lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.scheduler_classes[0] __snake_case : Dict = self.get_scheduler_config() __snake_case : Optional[int] = scheduler_class(**lowerCamelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(420 , 400 ) - 0.1_47_71 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(980 , 960 ) - 0.3_24_60 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(487 , 486 ) - 0.0_09_79 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(999 , 998 ) - 0.02 ) ) < 1E-5 def __snake_case ( self : Dict ) -> List[Any]: __snake_case : Any = self.scheduler_classes[0] __snake_case : Dict = self.get_scheduler_config() __snake_case : Optional[int] = scheduler_class(**lowerCamelCase ) __snake_case , __snake_case : str = 10, 0.0 scheduler.set_timesteps(lowerCamelCase ) __snake_case : Dict = self.dummy_model() __snake_case : List[Any] = self.dummy_sample_deter __snake_case : int = self.dummy_sample_deter + 0.1 __snake_case : str = self.dummy_sample_deter - 0.1 __snake_case : Dict = samplea.shape[0] __snake_case : str = torch.stack([samplea, samplea, samplea] , dim=0 ) __snake_case : int = torch.arange(lowerCamelCase )[0:3, None].repeat(1 , lowerCamelCase ) __snake_case : Union[str, Any] = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) __snake_case : Any = scheduler.batch_step_no_noise(lowerCamelCase , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) , lowerCamelCase ) __snake_case : Optional[Any] = torch.sum(torch.abs(lowerCamelCase ) ) __snake_case : List[Any] = torch.mean(torch.abs(lowerCamelCase ) ) assert abs(result_sum.item() - 11_47.79_04 ) < 1E-2 assert abs(result_mean.item() - 0.49_82 ) < 1E-3 def __snake_case ( self : str ) -> Optional[int]: __snake_case : Union[str, Any] = self.full_loop() __snake_case : Dict = torch.sum(torch.abs(lowerCamelCase ) ) __snake_case : Dict = torch.mean(torch.abs(lowerCamelCase ) ) assert abs(result_sum.item() - 1_72.00_67 ) < 1E-2 assert abs(result_mean.item() - 0.22_39_67 ) < 1E-3 def __snake_case ( self : str ) -> Dict: __snake_case : Any = self.full_loop(prediction_type="v_prediction" ) __snake_case : Optional[int] = torch.sum(torch.abs(lowerCamelCase ) ) __snake_case : int = torch.mean(torch.abs(lowerCamelCase ) ) assert abs(result_sum.item() - 52.53_02 ) < 1E-2 assert abs(result_mean.item() - 0.06_84 ) < 1E-3 def __snake_case ( self : Dict ) -> Tuple: # We specify different beta, so that the first alpha is 0.99 __snake_case : str = self.full_loop(set_alpha_to_one=lowerCamelCase , beta_start=0.01 ) __snake_case : Dict = torch.sum(torch.abs(lowerCamelCase ) ) __snake_case : Dict = torch.mean(torch.abs(lowerCamelCase ) ) assert abs(result_sum.item() - 1_49.82_95 ) < 1E-2 assert abs(result_mean.item() - 0.19_51 ) < 1E-3 def __snake_case ( self : Optional[Any] ) -> List[Any]: # We specify different beta, so that the first alpha is 0.99 __snake_case : Union[str, Any] = self.full_loop(set_alpha_to_one=lowerCamelCase , beta_start=0.01 ) __snake_case : Optional[Any] = torch.sum(torch.abs(lowerCamelCase ) ) __snake_case : Dict = torch.mean(torch.abs(lowerCamelCase ) ) assert abs(result_sum.item() - 1_49.07_84 ) < 1E-2 assert abs(result_mean.item() - 0.19_41 ) < 1E-3
81
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
import pandas as pd from matplotlib import pyplot as plt from sklearn.linear_model import LinearRegression # Splitting the dataset into the Training set and Test set from sklearn.model_selection import train_test_split # Fitting Polynomial Regression to the dataset from sklearn.preprocessing import PolynomialFeatures # Importing the dataset _snake_case : Union[str, Any] = pd.read_csv( "https://s3.us-west-2.amazonaws.com/public.gamelab.fun/dataset/" "position_salaries.csv" ) _snake_case : Optional[int] = dataset.iloc[:, 1:2].values _snake_case : int = dataset.iloc[:, 2].values _snake_case , _snake_case , _snake_case , _snake_case : Union[str, Any] = train_test_split(X, y, test_size=0.2, random_state=0) _snake_case : int = PolynomialFeatures(degree=4) _snake_case : str = poly_reg.fit_transform(X) _snake_case : Any = LinearRegression() pol_reg.fit(X_poly, y) def lowerCAmelCase_ ( ): plt.scatter(__lowerCamelCase , __lowerCamelCase , color="red" ) plt.plot(__lowerCamelCase , pol_reg.predict(poly_reg.fit_transform(__lowerCamelCase ) ) , color="blue" ) plt.title("Truth or Bluff (Linear Regression)" ) plt.xlabel("Position level" ) plt.ylabel("Salary" ) plt.show() if __name__ == "__main__": viz_polymonial() # Predicting a new result with Polymonial Regression pol_reg.predict(poly_reg.fit_transform([[5.5]])) # output should be 132148.43750003
81
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
1
from random import shuffle import tensorflow as tf from numpy import array def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Union[str, Any] = int(__lowerCamelCase ) assert noofclusters < len(__lowerCamelCase ) # Find out the dimensionality __snake_case : int = len(vectors[0] ) # Will help select random centroids from among the available vectors __snake_case : int = list(range(len(__lowerCamelCase ) ) ) shuffle(__lowerCamelCase ) # GRAPH OF COMPUTATION # We initialize a new graph and set it as the default during each run # of this algorithm. This ensures that as this function is called # multiple times, the default graph doesn't keep getting crowded with # unused ops and Variables from previous function calls. __snake_case : List[str] = tf.Graph() with graph.as_default(): # SESSION OF COMPUTATION __snake_case : Tuple = tf.Session() ##CONSTRUCTING THE ELEMENTS OF COMPUTATION ##First lets ensure we have a Variable vector for each centroid, ##initialized to one of the vectors from the available data points __snake_case : Tuple = [ tf.Variable(vectors[vector_indices[i]] ) for i in range(__lowerCamelCase ) ] ##These nodes will assign the centroid Variables the appropriate ##values __snake_case : List[str] = tf.placeholder("float64" , [dim] ) __snake_case : Optional[int] = [] for centroid in centroids: cent_assigns.append(tf.assign(__lowerCamelCase , __lowerCamelCase ) ) ##Variables for cluster assignments of individual vectors(initialized ##to 0 at first) __snake_case : Optional[int] = [tf.Variable(0 ) for i in range(len(__lowerCamelCase ) )] ##These nodes will assign an assignment Variable the appropriate ##value __snake_case : Union[str, Any] = tf.placeholder("int32" ) __snake_case : Any = [] for assignment in assignments: cluster_assigns.append(tf.assign(__lowerCamelCase , __lowerCamelCase ) ) ##Now lets construct the node that will compute the mean # The placeholder for the input __snake_case : Any = tf.placeholder("float" , [None, dim] ) # The Node/op takes the input and computes a mean along the 0th # dimension, i.e. the list of input vectors __snake_case : str = tf.reduce_mean(__lowerCamelCase , 0 ) ##Node for computing Euclidean distances # Placeholders for input __snake_case : Optional[Any] = tf.placeholder("float" , [dim] ) __snake_case : Tuple = tf.placeholder("float" , [dim] ) __snake_case : Tuple = tf.sqrt(tf.reduce_sum(tf.pow(tf.sub(__lowerCamelCase , __lowerCamelCase ) , 2 ) ) ) ##This node will figure out which cluster to assign a vector to, ##based on Euclidean distances of the vector from the centroids. # Placeholder for input __snake_case : Optional[Any] = tf.placeholder("float" , [noofclusters] ) __snake_case : List[str] = tf.argmin(__lowerCamelCase , 0 ) ##INITIALIZING STATE VARIABLES ##This will help initialization of all Variables defined with respect ##to the graph. The Variable-initializer should be defined after ##all the Variables have been constructed, so that each of them ##will be included in the initialization. __snake_case : Union[str, Any] = tf.initialize_all_variables() # Initialize all variables sess.run(__lowerCamelCase ) ##CLUSTERING ITERATIONS # Now perform the Expectation-Maximization steps of K-Means clustering # iterations. To keep things simple, we will only do a set number of # iterations, instead of using a Stopping Criterion. __snake_case : Any = 1_0_0 for _ in range(__lowerCamelCase ): ##EXPECTATION STEP ##Based on the centroid locations till last iteration, compute ##the _expected_ centroid assignments. # Iterate over each vector for vector_n in range(len(__lowerCamelCase ) ): __snake_case : Optional[int] = vectors[vector_n] # Compute Euclidean distance between this vector and each # centroid. Remember that this list cannot be named #'centroid_distances', since that is the input to the # cluster assignment node. __snake_case : Optional[int] = [ sess.run(__lowerCamelCase , feed_dict={va: vect, va: sess.run(__lowerCamelCase )} ) for centroid in centroids ] # Now use the cluster assignment node, with the distances # as the input __snake_case : Tuple = sess.run( __lowerCamelCase , feed_dict={centroid_distances: distances} ) # Now assign the value to the appropriate state variable sess.run( cluster_assigns[vector_n] , feed_dict={assignment_value: assignment} ) ##MAXIMIZATION STEP # Based on the expected state computed from the Expectation Step, # compute the locations of the centroids so as to maximize the # overall objective of minimizing within-cluster Sum-of-Squares for cluster_n in range(__lowerCamelCase ): # Collect all the vectors assigned to this cluster __snake_case : Dict = [ vectors[i] for i in range(len(__lowerCamelCase ) ) if sess.run(assignments[i] ) == cluster_n ] # Compute new centroid location __snake_case : List[str] = sess.run( __lowerCamelCase , feed_dict={mean_input: array(__lowerCamelCase )} ) # Assign value to appropriate variable sess.run( cent_assigns[cluster_n] , feed_dict={centroid_value: new_location} ) # Return centroids and assignments __snake_case : Any = sess.run(__lowerCamelCase ) __snake_case : Union[str, Any] = sess.run(__lowerCamelCase ) return centroids, assignments
81
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): while second != 0: __snake_case : List[str] = first & second first ^= second __snake_case : Tuple = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() _snake_case : List[Any] = int(input("Enter the first number: ").strip()) _snake_case : Optional[int] = int(input("Enter the second number: ").strip()) print(f'''{add(first, second) = }''')
81
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def lowerCAmelCase_ ( *__lowerCamelCase ): if not isinstance(__lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = list(__lowerCamelCase ) for i in range(len(__lowerCamelCase ) ): __snake_case : int = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = [ "CUDA out of memory.", # CUDA OOM "cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.", # CUDNN SNAFU "DefaultCPUAllocator: can't allocate memory", # CPU OOM ] if isinstance(__lowerCamelCase , __lowerCamelCase ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def lowerCAmelCase_ ( __lowerCamelCase = None , __lowerCamelCase = 1_2_8 ): if function is None: return functools.partial(__lowerCamelCase , starting_batch_size=__lowerCamelCase ) __snake_case : int = starting_batch_size def decorator(*__lowerCamelCase , **__lowerCamelCase ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() __snake_case : Optional[int] = list(inspect.signature(__lowerCamelCase ).parameters.keys() ) # Guard against user error if len(__lowerCamelCase ) < (len(__lowerCamelCase ) + 1): __snake_case : Optional[int] = ", ".join([F'{arg}={value}' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'Batch size was passed into `{function.__name__}` as the first argument when called.' F'Remove this as the decorator already does so: `{function.__name__}({arg_str})`' ) while True: if batch_size == 0: raise RuntimeError("No executable batch size found, reached zero." ) try: return function(__lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) except Exception as e: if should_reduce_batch_size(__lowerCamelCase ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
81
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
1
from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar _snake_case : Union[str, Any] = TypeVar("T") class a (Generic[T] ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list[T] , lowerCamelCase : Callable[[T, T], T] ) -> None: __snake_case : Any | T = None __snake_case : int = len(lowerCamelCase ) __snake_case : list[T] = [any_type for _ in range(self.N )] + arr __snake_case : Tuple = fnc self.build() def __snake_case ( self : Dict ) -> None: for p in range(self.N - 1 , 0 , -1 ): __snake_case : List[str] = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def __snake_case ( self : Optional[Any] , lowerCamelCase : int , lowerCamelCase : T ) -> None: p += self.N __snake_case : str = v while p > 1: __snake_case : Dict = p // 2 __snake_case : Optional[Any] = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def __snake_case ( self : List[str] , lowerCamelCase : int , lowerCamelCase : int ) -> T | None: # noqa: E741 __snake_case , __snake_case : Optional[int] = l + self.N, r + self.N __snake_case : T | None = None while l <= r: if l % 2 == 1: __snake_case : Optional[Any] = self.st[l] if res is None else self.fn(lowerCamelCase , self.st[l] ) if r % 2 == 0: __snake_case : int = self.st[r] if res is None else self.fn(lowerCamelCase , self.st[r] ) __snake_case , __snake_case : List[str] = (l + 1) // 2, (r - 1) // 2 return res if __name__ == "__main__": from functools import reduce _snake_case : Union[str, Any] = [1, 10, -2, 9, -3, 8, 4, -7, 5, 6, 11, -12] _snake_case : Dict = { 0: 7, 1: 2, 2: 6, 3: -14, 4: 5, 5: 4, 6: 7, 7: -10, 8: 9, 9: 10, 10: 12, 11: 1, } _snake_case : Optional[Any] = SegmentTree(test_array, min) _snake_case : Optional[Any] = SegmentTree(test_array, max) _snake_case : Optional[Any] = SegmentTree(test_array, lambda a, b: a + b) def lowerCAmelCase_ ( ): for i in range(len(__lowerCamelCase ) ): for j in range(__lowerCamelCase , len(__lowerCamelCase ) ): __snake_case : Optional[int] = reduce(__lowerCamelCase , test_array[i : j + 1] ) __snake_case : List[Any] = reduce(__lowerCamelCase , test_array[i : j + 1] ) __snake_case : Union[str, Any] = reduce(lambda __lowerCamelCase , __lowerCamelCase : a + b , test_array[i : j + 1] ) assert min_range == min_segment_tree.query(__lowerCamelCase , __lowerCamelCase ) assert max_range == max_segment_tree.query(__lowerCamelCase , __lowerCamelCase ) assert sum_range == sum_segment_tree.query(__lowerCamelCase , __lowerCamelCase ) test_all_segments() for index, value in test_updates.items(): _snake_case : List[Any] = value min_segment_tree.update(index, value) max_segment_tree.update(index, value) sum_segment_tree.update(index, value) test_all_segments()
81
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
1
def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Union[str, Any] = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): __snake_case : Union[str, Any] = n - k # Calculate C(n,k) for i in range(__lowerCamelCase ): result *= n - i result //= i + 1 return result def lowerCAmelCase_ ( __lowerCamelCase ): return binomial_coefficient(2 * node_count , __lowerCamelCase ) // (node_count + 1) def lowerCAmelCase_ ( __lowerCamelCase ): if n < 0: raise ValueError("factorial() not defined for negative values" ) __snake_case : Optional[int] = 1 for i in range(1 , n + 1 ): result *= i return result def lowerCAmelCase_ ( __lowerCamelCase ): return catalan_number(__lowerCamelCase ) * factorial(__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = int(input("Enter the number of nodes: ").strip() or 0) if node_count <= 0: raise ValueError("We need some nodes to work with.") print( f'''Given {node_count} nodes, there are {binary_tree_count(node_count)} ''' f'''binary trees and {catalan_number(node_count)} binary search trees.''' )
81
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : List[str] = { "configuration_mgp_str": ["MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP", "MgpstrConfig"], "processing_mgp_str": ["MgpstrProcessor"], "tokenization_mgp_str": ["MgpstrTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Union[str, Any] = [ "MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST", "MgpstrModel", "MgpstrPreTrainedModel", "MgpstrForSceneTextRecognition", ] if TYPE_CHECKING: from .configuration_mgp_str import MGP_STR_PRETRAINED_CONFIG_ARCHIVE_MAP, MgpstrConfig from .processing_mgp_str import MgpstrProcessor from .tokenization_mgp_str import MgpstrTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mgp_str import ( MGP_STR_PRETRAINED_MODEL_ARCHIVE_LIST, MgpstrForSceneTextRecognition, MgpstrModel, MgpstrPreTrainedModel, ) else: import sys _snake_case : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
1
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = DiTPipeline __UpperCAmelCase : Union[str, Any] = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS __UpperCAmelCase : List[Any] = PipelineTesterMixin.required_optional_params - { "latents", "num_images_per_prompt", "callback", "callback_steps", } __UpperCAmelCase : Optional[int] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS __UpperCAmelCase : Tuple = False def __snake_case ( self : int ) -> Any: torch.manual_seed(0 ) __snake_case : Dict = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=lowerCamelCase , activation_fn="gelu-approximate" , num_embeds_ada_norm=1000 , norm_type="ada_norm_zero" , norm_elementwise_affine=lowerCamelCase , ) __snake_case : Optional[int] = AutoencoderKL() __snake_case : str = DDIMScheduler() __snake_case : Union[str, Any] = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def __snake_case ( self : List[str] , lowerCamelCase : Any , lowerCamelCase : List[Any]=0 ) -> List[Any]: if str(lowerCamelCase ).startswith("mps" ): __snake_case : Optional[int] = torch.manual_seed(lowerCamelCase ) else: __snake_case : int = torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) __snake_case : Tuple = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def __snake_case ( self : List[Any] ) -> List[Any]: __snake_case : Dict = "cpu" __snake_case : Optional[int] = self.get_dummy_components() __snake_case : Dict = self.pipeline_class(**lowerCamelCase ) pipe.to(lowerCamelCase ) pipe.set_progress_bar_config(disable=lowerCamelCase ) __snake_case : Optional[Any] = self.get_dummy_inputs(lowerCamelCase ) __snake_case : Tuple = pipe(**lowerCamelCase ).images __snake_case : List[str] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) __snake_case : int = np.array([0.29_46, 0.66_01, 0.43_29, 0.32_96, 0.41_44, 0.53_19, 0.72_73, 0.50_13, 0.44_57] ) __snake_case : int = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(lowerCamelCase , 1E-3 ) def __snake_case ( self : Any ) -> Optional[Any]: self._test_inference_batch_single_identical(relax_max_difference=lowerCamelCase , expected_max_diff=1E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __snake_case ( self : Optional[Any] ) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @require_torch_gpu @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Optional[Any] ) -> Tuple: super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : List[Any] ) -> List[Any]: __snake_case : Union[str, Any] = torch.manual_seed(0 ) __snake_case : Any = DiTPipeline.from_pretrained("facebook/DiT-XL-2-256" ) pipe.to("cuda" ) __snake_case : Dict = ["vase", "umbrella", "white shark", "white wolf"] __snake_case : Any = pipe.get_label_ids(lowerCamelCase ) __snake_case : Tuple = pipe(lowerCamelCase , generator=lowerCamelCase , num_inference_steps=40 , output_type="np" ).images for word, image in zip(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = load_numpy( F'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy' ) assert np.abs((expected_image - image).max() ) < 1E-2 def __snake_case ( self : Dict ) -> Optional[int]: __snake_case : List[str] = DiTPipeline.from_pretrained("facebook/DiT-XL-2-512" ) __snake_case : Dict = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("cuda" ) __snake_case : List[str] = ["vase", "umbrella"] __snake_case : int = pipe.get_label_ids(lowerCamelCase ) __snake_case : Optional[Any] = torch.manual_seed(0 ) __snake_case : Dict = pipe(lowerCamelCase , generator=lowerCamelCase , num_inference_steps=25 , output_type="np" ).images for word, image in zip(lowerCamelCase , lowerCamelCase ): __snake_case : Tuple = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" F'/dit/{word}_512.npy' ) assert np.abs((expected_image - image).max() ) < 1E-1
81
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
1
import qiskit def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : int = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register __snake_case : Optional[Any] = qiskit.QuantumCircuit(__lowerCamelCase , __lowerCamelCase ) # Apply X (NOT) Gate to Qubits 0 & 1 circuit.x(0 ) circuit.x(1 ) # Map the quantum measurement to the classical bits circuit.measure([0, 1] , [0, 1] ) # Execute the circuit on the qasm simulator __snake_case : List[Any] = qiskit.execute(__lowerCamelCase , __lowerCamelCase , shots=1_0_0_0 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(__lowerCamelCase ) if __name__ == "__main__": _snake_case : int = single_qubit_measure(2, 2) print(f'''Total count for various states are: {counts}''')
81
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
1
# Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best hparam into the config # - generate model_cards - useful if you have multiple models from the same paper import argparse import json import os import re from collections import OrderedDict from os.path import basename, dirname import fairseq import torch from fairseq import hub_utils from fairseq.data.dictionary import Dictionary from transformers import FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() _snake_case : Any = 2 # based on the results of a search on a range of `num_beams`, `length_penalty` and `early_stopping` # values against wmt19 test data to obtain the best BLEU scores, we will use the following defaults: # # * `num_beams`: 5 (higher scores better, but requires more memory/is slower, can be adjusted by users) # * `early_stopping`: `False` consistently scored better # * `length_penalty` varied, so will assign the best one depending on the model _snake_case : List[str] = { # fairseq: "wmt19-ru-en": {"length_penalty": 1.1}, "wmt19-en-ru": {"length_penalty": 1.15}, "wmt19-en-de": {"length_penalty": 1.0}, "wmt19-de-en": {"length_penalty": 1.1}, # allenai: "wmt16-en-de-dist-12-1": {"length_penalty": 0.6}, "wmt16-en-de-dist-6-1": {"length_penalty": 0.6}, "wmt16-en-de-12-1": {"length_penalty": 0.8}, "wmt19-de-en-6-6-base": {"length_penalty": 0.6}, "wmt19-de-en-6-6-big": {"length_penalty": 0.6}, } # this remaps the different models to their organization names _snake_case : int = {} for m in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: _snake_case : Optional[Any] = "facebook" for m in [ "wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1", "wmt19-de-en-6-6-base", "wmt19-de-en-6-6-big", ]: _snake_case : Optional[Any] = "allenai" def lowerCAmelCase_ ( __lowerCamelCase ): # (1) remove word breaking symbol, (2) add word ending symbol where the word is not broken up, # e.g.: d = {'le@@': 5, 'tt@@': 6, 'er': 7} => {'le': 5, 'tt': 6, 'er</w>': 7} __snake_case : List[Any] = dict((re.sub(R"@@$" , "" , __lowerCamelCase ), v) if k.endswith("@@" ) else (re.sub(R"$" , "</w>" , __lowerCamelCase ), v) for k, v in d.items() ) __snake_case : Tuple = "<s> <pad> </s> <unk>".split() # restore the special tokens for k in keep_keys: del da[F'{k}</w>'] __snake_case : Optional[int] = d[k] # restore return da def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): # prep assert os.path.exists(__lowerCamelCase ) os.makedirs(__lowerCamelCase , exist_ok=__lowerCamelCase ) print(F'Writing results to {pytorch_dump_folder_path}' ) # handle various types of models __snake_case : Any = basename(__lowerCamelCase ) __snake_case : Union[str, Any] = dirname(__lowerCamelCase ) __snake_case : int = fairseq.model_parallel.models.transformer.ModelParallelTransformerModel __snake_case : Optional[Any] = cls.hub_models() __snake_case : Union[str, Any] = {"bpe": "fastbpe", "tokenizer": "moses"} __snake_case : Any = "." # note: since the model dump is old, fairseq has upgraded its model some # time later, and it does a whole lot of rewrites and splits on the saved # weights, therefore we can't use torch.load() directly on the model file. # see: upgrade_state_dict(state_dict) in fairseq_model.py print(F'using checkpoint {checkpoint_file}' ) __snake_case : Dict = hub_utils.from_pretrained( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , archive_map=__lowerCamelCase , **__lowerCamelCase ) __snake_case : List[Any] = vars(chkpt["args"]["model"] ) __snake_case : int = args["source_lang"] __snake_case : str = args["target_lang"] __snake_case : List[str] = dirname(__lowerCamelCase ) __snake_case : str = basename(__lowerCamelCase ) # dicts __snake_case : str = os.path.join(__lowerCamelCase , F'dict.{src_lang}.txt' ) __snake_case : List[str] = os.path.join(__lowerCamelCase , F'dict.{tgt_lang}.txt' ) __snake_case : List[str] = Dictionary.load(__lowerCamelCase ) __snake_case : Any = rewrite_dict_keys(src_dict.indices ) __snake_case : str = len(__lowerCamelCase ) __snake_case : List[str] = os.path.join(__lowerCamelCase , "vocab-src.json" ) print(F'Generating {src_vocab_file} of {src_vocab_size} of {src_lang} records' ) with open(__lowerCamelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # detect whether this is a do_lower_case situation, which can be derived by checking whether we # have at least one uppercase letter in the source vocab __snake_case : Tuple = True for k in src_vocab.keys(): if not k.islower(): __snake_case : Tuple = False break __snake_case : Dict = Dictionary.load(__lowerCamelCase ) __snake_case : List[Any] = rewrite_dict_keys(tgt_dict.indices ) __snake_case : Any = len(__lowerCamelCase ) __snake_case : List[str] = os.path.join(__lowerCamelCase , "vocab-tgt.json" ) print(F'Generating {tgt_vocab_file} of {tgt_vocab_size} of {tgt_lang} records' ) with open(__lowerCamelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # merges_file (bpecodes) __snake_case : Union[str, Any] = os.path.join(__lowerCamelCase , VOCAB_FILES_NAMES["merges_file"] ) for fn in ["bpecodes", "code"]: # older fairseq called the merges file "code" __snake_case : Any = os.path.join(__lowerCamelCase , __lowerCamelCase ) if os.path.exists(__lowerCamelCase ): break with open(__lowerCamelCase , encoding="utf-8" ) as fin: __snake_case : int = fin.read() __snake_case : Any = re.sub(R" \d+$" , "" , __lowerCamelCase , 0 , re.M ) # remove frequency number print(F'Generating {merges_file}' ) with open(__lowerCamelCase , "w" , encoding="utf-8" ) as fout: fout.write(__lowerCamelCase ) # model config __snake_case : str = os.path.join(__lowerCamelCase , "config.json" ) # validate bpe/tokenizer config, as currently it's hardcoded to moses+fastbpe - # may have to modify the tokenizer if a different type is used by a future model assert args["bpe"] == "fastbpe", F'need to extend tokenizer to support bpe={args["bpe"]}' assert args["tokenizer"] == "moses", F'need to extend tokenizer to support bpe={args["tokenizer"]}' __snake_case : List[Any] = { "architectures": ["FSMTForConditionalGeneration"], "model_type": "fsmt", "activation_dropout": args["activation_dropout"], "activation_function": "relu", "attention_dropout": args["attention_dropout"], "d_model": args["decoder_embed_dim"], "dropout": args["dropout"], "init_std": 0.0_2, "max_position_embeddings": args["max_source_positions"], "num_hidden_layers": args["encoder_layers"], "src_vocab_size": src_vocab_size, "tgt_vocab_size": tgt_vocab_size, "langs": [src_lang, tgt_lang], "encoder_attention_heads": args["encoder_attention_heads"], "encoder_ffn_dim": args["encoder_ffn_embed_dim"], "encoder_layerdrop": args["encoder_layerdrop"], "encoder_layers": args["encoder_layers"], "decoder_attention_heads": args["decoder_attention_heads"], "decoder_ffn_dim": args["decoder_ffn_embed_dim"], "decoder_layerdrop": args["decoder_layerdrop"], "decoder_layers": args["decoder_layers"], "bos_token_id": 0, "pad_token_id": 1, "eos_token_id": 2, "is_encoder_decoder": True, "scale_embedding": not args["no_scale_embedding"], "tie_word_embeddings": args["share_all_embeddings"], } # good hparam defaults to start with __snake_case : Optional[Any] = 5 __snake_case : List[Any] = False if model_dir in best_score_hparams and "length_penalty" in best_score_hparams[model_dir]: __snake_case : List[str] = best_score_hparams[model_dir]["length_penalty"] else: __snake_case : Tuple = 1.0 print(F'Generating {fsmt_model_config_file}' ) with open(__lowerCamelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # tokenizer config __snake_case : List[str] = os.path.join(__lowerCamelCase , __lowerCamelCase ) __snake_case : Tuple = { "langs": [src_lang, tgt_lang], "model_max_length": 1_0_2_4, "do_lower_case": do_lower_case, } print(F'Generating {fsmt_tokenizer_config_file}' ) with open(__lowerCamelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(__lowerCamelCase , ensure_ascii=__lowerCamelCase , indent=__lowerCamelCase ) ) # model __snake_case : List[Any] = chkpt["models"][0] __snake_case : Tuple = model.state_dict() # rename keys to start with 'model.' __snake_case : Union[str, Any] = OrderedDict(("model." + k, v) for k, v in model_state_dict.items() ) # remove unneeded keys __snake_case : Any = [ "model.model", "model.encoder.version", "model.decoder.version", "model.encoder_embed_tokens.weight", "model.decoder_embed_tokens.weight", "model.encoder.embed_positions._float_tensor", "model.decoder.embed_positions._float_tensor", ] for k in ignore_keys: model_state_dict.pop(__lowerCamelCase , __lowerCamelCase ) __snake_case : Union[str, Any] = FSMTConfig.from_pretrained(__lowerCamelCase ) __snake_case : Union[str, Any] = FSMTForConditionalGeneration(__lowerCamelCase ) # check that it loads ok model_new.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) # save __snake_case : int = os.path.join(__lowerCamelCase , __lowerCamelCase ) print(F'Generating {pytorch_weights_dump_path}' ) torch.save(__lowerCamelCase , __lowerCamelCase ) print("Conversion is done!" ) print("\nLast step is to upload the files to s3" ) print(F'cd {data_root}' ) print(F'transformers-cli upload {model_dir}' ) if __name__ == "__main__": _snake_case : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--fsmt_checkpoint_path", default=None, type=str, required=True, help=( "Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts," " bpecodes, etc." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _snake_case : Any = parser.parse_args() convert_fsmt_checkpoint_to_pytorch(args.fsmt_checkpoint_path, args.pytorch_dump_folder_path)
81
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available _snake_case : Optional[Any] = { "configuration_tapas": ["TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP", "TapasConfig"], "tokenization_tapas": ["TapasTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Tuple = [ "TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST", "TapasForMaskedLM", "TapasForQuestionAnswering", "TapasForSequenceClassification", "TapasModel", "TapasPreTrainedModel", "load_tf_weights_in_tapas", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ "TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST", "TFTapasForMaskedLM", "TFTapasForQuestionAnswering", "TFTapasForSequenceClassification", "TFTapasModel", "TFTapasPreTrainedModel", ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys _snake_case : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
import functools def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = len(__lowerCamelCase ) __snake_case : List[Any] = len(__lowerCamelCase ) @functools.cache def min_distance(__lowerCamelCase , __lowerCamelCase ) -> int: # if first word index is overflow - delete all from the second word if indexa >= len_worda: return len_worda - indexa # if second word index is overflow - delete all from the first word if indexa >= len_worda: return len_worda - indexa __snake_case : List[str] = int(worda[indexa] != worda[indexa] ) # current letters not identical return min( 1 + min_distance(indexa + 1 , __lowerCamelCase ) , 1 + min_distance(__lowerCamelCase , indexa + 1 ) , diff + min_distance(indexa + 1 , indexa + 1 ) , ) return min_distance(0 , 0 ) if __name__ == "__main__": import doctest doctest.testmod()
81
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
1
import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a : """simple docstring""" def __init__( self : int , lowerCamelCase : Optional[Any] , lowerCamelCase : Union[str, Any]=13 , lowerCamelCase : Any=[30, 30] , lowerCamelCase : Optional[Any]=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : str=True , lowerCamelCase : List[str]=True , lowerCamelCase : str=32 , lowerCamelCase : Optional[Any]=5 , lowerCamelCase : str=4 , lowerCamelCase : Dict=37 , lowerCamelCase : List[str]="gelu" , lowerCamelCase : List[str]=0.1 , lowerCamelCase : Tuple=0.1 , lowerCamelCase : int=10 , lowerCamelCase : Union[str, Any]=0.02 , lowerCamelCase : Optional[Any]=3 , lowerCamelCase : Optional[Any]=None , lowerCamelCase : List[str]=8 , lowerCamelCase : List[Any]=10 , ) -> Tuple: __snake_case : Dict = parent __snake_case : Tuple = batch_size __snake_case : str = image_size __snake_case : Tuple = patch_size __snake_case : str = num_channels __snake_case : int = is_training __snake_case : int = use_labels __snake_case : Dict = hidden_size __snake_case : List[Any] = num_hidden_layers __snake_case : List[Any] = num_attention_heads __snake_case : Any = intermediate_size __snake_case : Tuple = hidden_act __snake_case : Optional[int] = hidden_dropout_prob __snake_case : int = attention_probs_dropout_prob __snake_case : Optional[Any] = type_sequence_label_size __snake_case : int = initializer_range __snake_case : Optional[Any] = num_labels __snake_case : List[str] = scope __snake_case : Dict = n_targets __snake_case : Dict = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens __snake_case : int = (image_size[1] // patch_size) * (image_size[0] // patch_size) __snake_case : Optional[Any] = num_patches + 1 + self.num_detection_tokens def __snake_case ( self : Tuple ) -> List[Any]: __snake_case : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) __snake_case : str = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) __snake_case : Dict = [] for i in range(self.batch_size ): __snake_case : Optional[int] = {} __snake_case : List[Any] = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=lowerCamelCase ) __snake_case : List[str] = torch.rand(self.n_targets , 4 , device=lowerCamelCase ) labels.append(lowerCamelCase ) __snake_case : Union[str, Any] = self.get_config() return config, pixel_values, labels def __snake_case ( self : Dict ) -> str: return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def __snake_case ( self : Dict , lowerCamelCase : str , lowerCamelCase : Optional[Any] , lowerCamelCase : str ) -> Optional[Any]: __snake_case : int = YolosModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Dict = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def __snake_case ( self : Tuple , lowerCamelCase : int , lowerCamelCase : Union[str, Any] , lowerCamelCase : Optional[int] ) -> Any: __snake_case : List[str] = YolosForObjectDetection(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Optional[Any] = model(pixel_values=lowerCamelCase ) __snake_case : Dict = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) __snake_case : Optional[int] = model(pixel_values=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def __snake_case ( self : Dict ) -> Any: __snake_case : Tuple = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case : Dict = config_and_inputs __snake_case : str = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : int = (YolosModel, YolosForObjectDetection) if is_torch_available() else () __UpperCAmelCase : List[str] = ( {"feature-extraction": YolosModel, "object-detection": YolosForObjectDetection} if is_torch_available() else {} ) __UpperCAmelCase : List[Any] = False __UpperCAmelCase : Any = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[Any] = False def __snake_case ( self : Optional[int] , lowerCamelCase : int , lowerCamelCase : List[str] , lowerCamelCase : Union[str, Any]=False ) -> List[Any]: __snake_case : Union[str, Any] = super()._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": __snake_case : Dict = [] for i in range(self.model_tester.batch_size ): __snake_case : Tuple = {} __snake_case : Optional[Any] = torch.ones( size=(self.model_tester.n_targets,) , device=lowerCamelCase , dtype=torch.long ) __snake_case : Any = torch.ones( self.model_tester.n_targets , 4 , device=lowerCamelCase , dtype=torch.float ) labels.append(lowerCamelCase ) __snake_case : Dict = labels return inputs_dict def __snake_case ( self : Tuple ) -> Dict: __snake_case : Any = YolosModelTester(self ) __snake_case : str = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def __snake_case ( self : str ) -> Optional[Any]: self.config_tester.run_common_tests() def __snake_case ( self : Dict ) -> Optional[int]: # YOLOS does not use inputs_embeds pass def __snake_case ( self : int ) -> int: __snake_case , __snake_case : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __snake_case : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def __snake_case ( self : Any ) -> Optional[int]: __snake_case , __snake_case : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = model_class(lowerCamelCase ) __snake_case : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Any = [*signature.parameters.keys()] __snake_case : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : Tuple ) -> Optional[int]: __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : int = True # in YOLOS, the seq_len is different __snake_case : Union[str, Any] = self.model_tester.expected_seq_len for model_class in self.all_model_classes: __snake_case : List[Any] = True __snake_case : str = False __snake_case : str = True __snake_case : str = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : Optional[Any] = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : int = outputs.attentions self.assertEqual(len(lowerCamelCase ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __snake_case : int = True __snake_case : List[str] = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : str = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : str = outputs.attentions self.assertEqual(len(lowerCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) __snake_case : List[Any] = len(lowerCamelCase ) # Check attention is always last and order is fine __snake_case : Optional[int] = True __snake_case : str = True __snake_case : Dict = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : Optional[Any] = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Tuple = 1 self.assertEqual(out_len + added_hidden_states , len(lowerCamelCase ) ) __snake_case : Any = outputs.attentions self.assertEqual(len(lowerCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def __snake_case ( self : int ) -> Dict: def check_hidden_states_output(lowerCamelCase : Dict , lowerCamelCase : Union[str, Any] , lowerCamelCase : str ): __snake_case : Union[str, Any] = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : List[Any] = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : List[Any] = outputs.hidden_states __snake_case : str = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # YOLOS has a different seq_length __snake_case : Optional[Any] = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Any = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : Optional[int] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*lowerCamelCase ) @slow def __snake_case ( self : str ) -> int: for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[Any] = YolosModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : List[Any] ) -> Union[str, Any]: return AutoImageProcessor.from_pretrained("hustvl/yolos-small" ) if is_vision_available() else None @slow def __snake_case ( self : Any ) -> Optional[Any]: __snake_case : int = YolosForObjectDetection.from_pretrained("hustvl/yolos-small" ).to(lowerCamelCase ) __snake_case : Union[str, Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : Optional[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Optional[int] = model(inputs.pixel_values ) # verify outputs __snake_case : Dict = torch.Size((1, 100, 92) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] , device=lowerCamelCase , ) __snake_case : Optional[int] = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] , device=lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , lowerCamelCase , atol=1E-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , lowerCamelCase , atol=1E-4 ) ) # verify postprocessing __snake_case : Any = image_processor.post_process_object_detection( lowerCamelCase , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] __snake_case : List[Any] = torch.tensor([0.99_94, 0.97_90, 0.99_64, 0.99_72, 0.98_61] ).to(lowerCamelCase ) __snake_case : List[Any] = [75, 75, 17, 63, 17] __snake_case : int = torch.tensor([3_35.06_09, 79.38_48, 3_75.42_16, 1_87.24_95] ).to(lowerCamelCase ) self.assertEqual(len(results["scores"] ) , 5 ) self.assertTrue(torch.allclose(results["scores"] , lowerCamelCase , atol=1E-4 ) ) self.assertSequenceEqual(results["labels"].tolist() , lowerCamelCase ) self.assertTrue(torch.allclose(results["boxes"][0, :] , lowerCamelCase ) )
81
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Union[str, Any] = current_set.copy() for row_index, row in enumerate(__lowerCamelCase ): __snake_case : Any = row[0] for column_index, column in enumerate(__lowerCamelCase ): if magnitude == 0: __snake_case : Dict = column continue __snake_case : Optional[Any] = column / magnitude # Subtract to cancel term __snake_case : Any = current_set[0] __snake_case : Optional[Any] = [first_row] __snake_case : Optional[int] = current_set[1::] for row in current_set: __snake_case : Optional[Any] = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(__lowerCamelCase ) continue for column_index in range(len(__lowerCamelCase ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(__lowerCamelCase ) # Create next recursion iteration set if len(final_set[0] ) != 3: __snake_case : str = final_set[0] __snake_case : Union[str, Any] = [] __snake_case : List[Any] = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) __snake_case : List[str] = simplify(__lowerCamelCase ) for i in range(len(__lowerCamelCase ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , __lowerCamelCase ) __snake_case : Optional[Any] = resultant return final_set def lowerCAmelCase_ ( __lowerCamelCase ): if len(__lowerCamelCase ) == 0: raise IndexError("solve_simultaneous() requires n lists of length n+1" ) __snake_case : List[str] = len(__lowerCamelCase ) + 1 if any(len(__lowerCamelCase ) != _length for item in equations ): raise IndexError("solve_simultaneous() requires n lists of length n+1" ) for row in equations: if any(not isinstance(__lowerCamelCase , (int, float) ) for column in row ): raise ValueError("solve_simultaneous() requires lists of integers" ) if len(__lowerCamelCase ) == 1: return [equations[0][-1] / equations[0][0]] __snake_case : List[Any] = equations.copy() if any(0 in row for row in data_set ): __snake_case : str = data_set.copy() __snake_case : str = [] for row_index, row in enumerate(__lowerCamelCase ): if 0 not in row: __snake_case : Union[str, Any] = data_set.pop(__lowerCamelCase ) break if not full_row: raise ValueError("solve_simultaneous() requires at least 1 full equation" ) data_set.insert(0 , __lowerCamelCase ) __snake_case : Any = data_set.copy() __snake_case : Union[str, Any] = simplify(__lowerCamelCase ) __snake_case : List[Any] = simplified[::-1] __snake_case : list = [] for row in simplified: __snake_case : int = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue __snake_case : Optional[int] = row.copy()[: len(__lowerCamelCase ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(__lowerCamelCase ) == 0: solutions.append(0 ) continue __snake_case : List[Any] = temp_row[1::] __snake_case : str = temp_row[::-1] for column_index, column in enumerate(__lowerCamelCase ): current_solution -= column * solutions[column_index] solutions.append(__lowerCamelCase ) __snake_case : Dict = [] for item in solutions: final.append(float(round(__lowerCamelCase , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() _snake_case : Union[str, Any] = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
81
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
1
from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a : """simple docstring""" def __init__( self : Dict , lowerCamelCase : int , lowerCamelCase : int=13 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=2 , lowerCamelCase : int=3 , lowerCamelCase : List[str]=True , lowerCamelCase : Any=True , lowerCamelCase : List[Any]=32 , lowerCamelCase : List[str]=2 , lowerCamelCase : int=4 , lowerCamelCase : str=37 , lowerCamelCase : Optional[Any]="gelu" , lowerCamelCase : Tuple=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Tuple=10 , lowerCamelCase : Union[str, Any]=0.02 , lowerCamelCase : Any=3 , lowerCamelCase : Any=0.6 , lowerCamelCase : Optional[int]=None , ) -> Union[str, Any]: __snake_case : List[str] = parent __snake_case : List[str] = batch_size __snake_case : Optional[int] = image_size __snake_case : List[Any] = patch_size __snake_case : Union[str, Any] = num_channels __snake_case : Any = is_training __snake_case : List[str] = use_labels __snake_case : Optional[Any] = hidden_size __snake_case : int = num_hidden_layers __snake_case : List[str] = num_attention_heads __snake_case : Tuple = intermediate_size __snake_case : Tuple = hidden_act __snake_case : Dict = hidden_dropout_prob __snake_case : str = attention_probs_dropout_prob __snake_case : Any = type_sequence_label_size __snake_case : str = initializer_range __snake_case : int = mask_ratio __snake_case : int = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) __snake_case : Optional[int] = (image_size // patch_size) ** 2 __snake_case : Union[str, Any] = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def __snake_case ( self : Optional[int] ) -> str: __snake_case : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None if self.use_labels: __snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case : Tuple = self.get_config() return config, pixel_values, labels def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def __snake_case ( self : List[str] , lowerCamelCase : List[str] , lowerCamelCase : str , lowerCamelCase : Optional[int] ) -> List[Any]: __snake_case : List[Any] = TFViTMAEModel(config=lowerCamelCase ) __snake_case : Optional[Any] = model(lowerCamelCase , training=lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Any , lowerCamelCase : Dict , lowerCamelCase : int ) -> Union[str, Any]: __snake_case : Tuple = TFViTMAEForPreTraining(lowerCamelCase ) __snake_case : Tuple = model(lowerCamelCase , training=lowerCamelCase ) # expected sequence length = num_patches __snake_case : Union[str, Any] = (self.image_size // self.patch_size) ** 2 __snake_case : Union[str, Any] = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images __snake_case : Optional[int] = 1 __snake_case : Optional[int] = TFViTMAEForPreTraining(lowerCamelCase ) __snake_case : List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __snake_case : List[Any] = model(lowerCamelCase , training=lowerCamelCase ) __snake_case : Dict = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def __snake_case ( self : Union[str, Any] ) -> int: __snake_case : List[str] = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case)) : Optional[Any] = config_and_inputs __snake_case : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Dict = (TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () __UpperCAmelCase : List[Any] = {"feature-extraction": TFViTMAEModel} if is_tf_available() else {} __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : int = False __UpperCAmelCase : int = False def __snake_case ( self : int ) -> Tuple: __snake_case : Any = TFViTMAEModelTester(self ) __snake_case : List[str] = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def __snake_case ( self : Optional[Any] ) -> str: self.config_tester.run_common_tests() @unittest.skip(reason="ViTMAE does not use inputs_embeds" ) def __snake_case ( self : List[str] ) -> int: pass def __snake_case ( self : Optional[int] ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : int = model_class(lowerCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) __snake_case : Optional[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , tf.keras.layers.Layer ) ) def __snake_case ( self : int ) -> Tuple: __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : List[str] = model_class(lowerCamelCase ) __snake_case : str = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Tuple = [*signature.parameters.keys()] __snake_case : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> int: __snake_case : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Dict: # make the mask reproducible np.random.seed(2 ) __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : List[Any] = int((config.image_size // config.patch_size) ** 2 ) __snake_case : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __snake_case : Any = model_class(lowerCamelCase ) __snake_case : List[str] = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) __snake_case : int = model(lowerCamelCase , noise=lowerCamelCase ) __snake_case : List[Any] = copy.deepcopy(self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : List[str] = model(**lowerCamelCase , noise=lowerCamelCase ) __snake_case : str = outputs_dict[0].numpy() __snake_case : Tuple = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1E-6 ) def __snake_case ( self : Union[str, Any] ) -> List[str]: # make the mask reproducible np.random.seed(2 ) __snake_case , __snake_case : Dict = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Tuple = int((config.image_size // config.patch_size) ** 2 ) __snake_case : List[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(lowerCamelCase : Optional[int] ): __snake_case : Any = {} for k, v in inputs_dict.items(): if tf.is_tensor(lowerCamelCase ): __snake_case : Any = v.numpy() else: __snake_case : Optional[int] = np.array(lowerCamelCase ) return inputs_np_dict for model_class in self.all_model_classes: __snake_case : Optional[int] = model_class(lowerCamelCase ) __snake_case : List[str] = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) __snake_case : Any = prepare_numpy_arrays(lowerCamelCase ) __snake_case : Tuple = model(lowerCamelCase , noise=lowerCamelCase ) __snake_case : Tuple = model(**lowerCamelCase , noise=lowerCamelCase ) self.assert_outputs_same(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : str , lowerCamelCase : str , lowerCamelCase : List[str] , lowerCamelCase : Union[str, Any] ) -> Optional[Any]: # make masks reproducible np.random.seed(2 ) __snake_case : Optional[int] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) __snake_case : Tuple = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __snake_case : Tuple = tf.constant(lowerCamelCase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument __snake_case : Optional[int] = tf_noise super().check_pt_tf_models(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Any: # make mask reproducible np.random.seed(2 ) __snake_case , __snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : int = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(lowerCamelCase ) if module_member_name.endswith("MainLayer" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("MainLayer" )] == model_class.__name__[: -len("Model" )] for module_member in (getattr(lowerCamelCase , lowerCamelCase ),) if isinstance(lowerCamelCase , lowerCamelCase ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(lowerCamelCase , "_keras_serializable" , lowerCamelCase ) } __snake_case : List[str] = int((config.image_size // config.patch_size) ** 2 ) __snake_case : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) __snake_case : List[str] = tf.convert_to_tensor(lowerCamelCase ) inputs_dict.update({"noise": noise} ) for main_layer_class in tf_main_layer_classes: __snake_case : Optional[int] = main_layer_class(lowerCamelCase ) __snake_case : str = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } __snake_case : Optional[int] = tf.keras.Model(lowerCamelCase , outputs=main_layer(lowerCamelCase ) ) __snake_case : List[str] = model(lowerCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: __snake_case : List[Any] = os.path.join(lowerCamelCase , "keras_model.h5" ) model.save(lowerCamelCase ) __snake_case : Optional[int] = tf.keras.models.load_model( lowerCamelCase , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(lowerCamelCase , tf.keras.Model ) __snake_case : Any = model(lowerCamelCase ) self.assert_outputs_same(lowerCamelCase , lowerCamelCase ) @slow def __snake_case ( self : Dict ) -> int: # make mask reproducible np.random.seed(2 ) __snake_case , __snake_case : Dict = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Dict = int((config.image_size // config.patch_size) ** 2 ) __snake_case : Dict = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __snake_case : Any = model_class(lowerCamelCase ) __snake_case : Dict = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) __snake_case : str = model(lowerCamelCase , noise=lowerCamelCase ) if model_class.__name__ == "TFViTMAEModel": __snake_case : Tuple = outputs.last_hidden_state.numpy() __snake_case : Dict = 0 else: __snake_case : List[Any] = outputs.logits.numpy() __snake_case : Tuple = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowerCamelCase , saved_model=lowerCamelCase ) __snake_case : Union[str, Any] = model_class.from_pretrained(lowerCamelCase ) __snake_case : Dict = model(lowerCamelCase , noise=lowerCamelCase ) if model_class.__name__ == "TFViTMAEModel": __snake_case : Tuple = after_outputs["last_hidden_state"].numpy() __snake_case : List[Any] = 0 else: __snake_case : int = after_outputs["logits"].numpy() __snake_case : str = 0 __snake_case : List[str] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowerCamelCase , 1E-5 ) def __snake_case ( self : Optional[Any] ) -> Tuple: # make mask reproducible np.random.seed(2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Union[str, Any] = int((config.image_size // config.patch_size) ** 2 ) __snake_case : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: __snake_case : Tuple = model_class(lowerCamelCase ) __snake_case : List[Any] = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) __snake_case : Any = model(lowerCamelCase , noise=lowerCamelCase ) __snake_case : Union[str, Any] = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(lowerCamelCase ) __snake_case : Any = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config __snake_case : Optional[int] = model_class.from_config(model.config ) __snake_case : Any = new_model(lowerCamelCase ) # Build model new_model.set_weights(model.get_weights() ) __snake_case : str = new_model(lowerCamelCase , noise=lowerCamelCase ) self.assert_outputs_same(lowerCamelCase , lowerCamelCase ) @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def __snake_case ( self : Optional[int] ) -> Optional[Any]: pass @unittest.skip(reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load" ) def __snake_case ( self : Optional[int] ) -> List[Any]: pass @slow def __snake_case ( self : List[Any] ) -> Tuple: __snake_case : str = TFViTMAEModel.from_pretrained("google/vit-base-patch16-224" ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> str: return ViTImageProcessor.from_pretrained("facebook/vit-mae-base" ) if is_vision_available() else None @slow def __snake_case ( self : Tuple ) -> Tuple: # make random mask reproducible across the PT and TF model np.random.seed(2 ) __snake_case : Any = TFViTMAEForPreTraining.from_pretrained("facebook/vit-mae-base" ) __snake_case : Tuple = self.default_image_processor __snake_case : List[str] = prepare_img() __snake_case : Optional[Any] = image_processor(images=lowerCamelCase , return_tensors="tf" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) __snake_case : List[Any] = ViTMAEConfig() __snake_case : str = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) __snake_case : Tuple = np.random.uniform(size=(1, num_patches) ) # forward pass __snake_case : str = model(**lowerCamelCase , noise=lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[str] = tf.convert_to_tensor( [[-0.05_48, -1.70_23, -0.93_25], [0.37_21, -0.56_70, -0.22_33], [0.82_35, -1.38_78, -0.35_24]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , lowerCamelCase , atol=1E-4 )
81
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
1
def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : Dict = set(range(3 , __lowerCamelCase , 2 ) ) primes.add(2 ) for p in range(3 , __lowerCamelCase , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , __lowerCamelCase , __lowerCamelCase ) ) ) __snake_case : Any = [float(__lowerCamelCase ) for n in range(limit + 1 )] for p in primes: for n in range(__lowerCamelCase , limit + 1 , __lowerCamelCase ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'''{solution() = }''')
81
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
1
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class a : """simple docstring""" __UpperCAmelCase : CommonSchedulerState # setable values __UpperCAmelCase : jnp.ndarray __UpperCAmelCase : jnp.ndarray __UpperCAmelCase : Optional[int] = None @classmethod def __snake_case ( cls : str , lowerCamelCase : CommonSchedulerState , lowerCamelCase : jnp.ndarray , lowerCamelCase : jnp.ndarray ) -> int: return cls(common=lowerCamelCase , init_noise_sigma=lowerCamelCase , timesteps=lowerCamelCase ) @dataclass class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : DDPMSchedulerState class a (_lowerCAmelCase , _lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = [e.name for e in FlaxKarrasDiffusionSchedulers] __UpperCAmelCase : jnp.dtype @property def __snake_case ( self : str ) -> Optional[int]: return True @register_to_config def __init__( self : int , lowerCamelCase : int = 1000 , lowerCamelCase : float = 0.00_01 , lowerCamelCase : float = 0.02 , lowerCamelCase : str = "linear" , lowerCamelCase : Optional[jnp.ndarray] = None , lowerCamelCase : str = "fixed_small" , lowerCamelCase : bool = True , lowerCamelCase : str = "epsilon" , lowerCamelCase : jnp.dtype = jnp.floataa , ) -> Any: __snake_case : List[str] = dtype def __snake_case ( self : Dict , lowerCamelCase : Optional[CommonSchedulerState] = None ) -> DDPMSchedulerState: if common is None: __snake_case : Optional[int] = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution __snake_case : int = jnp.array(1.0 , dtype=self.dtype ) __snake_case : List[str] = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=lowerCamelCase , init_noise_sigma=lowerCamelCase , timesteps=lowerCamelCase , ) def __snake_case ( self : int , lowerCamelCase : DDPMSchedulerState , lowerCamelCase : jnp.ndarray , lowerCamelCase : Optional[int] = None ) -> jnp.ndarray: return sample def __snake_case ( self : Dict , lowerCamelCase : DDPMSchedulerState , lowerCamelCase : int , lowerCamelCase : Tuple = () ) -> DDPMSchedulerState: __snake_case : List[Any] = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 __snake_case : Tuple = (jnp.arange(0 , lowerCamelCase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=lowerCamelCase , timesteps=lowerCamelCase , ) def __snake_case ( self : int , lowerCamelCase : DDPMSchedulerState , lowerCamelCase : List[Any] , lowerCamelCase : List[str]=None , lowerCamelCase : Optional[int]=None ) -> Optional[int]: __snake_case : Union[str, Any] = state.common.alphas_cumprod[t] __snake_case : Union[str, Any] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample __snake_case : List[Any] = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: __snake_case : int = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": __snake_case : Tuple = jnp.clip(lowerCamelCase , a_min=1E-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": __snake_case : List[str] = jnp.log(jnp.clip(lowerCamelCase , a_min=1E-20 ) ) elif variance_type == "fixed_large": __snake_case : Union[str, Any] = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log __snake_case : Dict = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": __snake_case : List[Any] = variance __snake_case : Any = state.common.betas[t] __snake_case : Any = (predicted_variance + 1) / 2 __snake_case : Tuple = frac * max_log + (1 - frac) * min_log return variance def __snake_case ( self : Any , lowerCamelCase : DDPMSchedulerState , lowerCamelCase : jnp.ndarray , lowerCamelCase : int , lowerCamelCase : jnp.ndarray , lowerCamelCase : Optional[jax.random.KeyArray] = None , lowerCamelCase : bool = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: __snake_case : int = timestep if key is None: __snake_case : int = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: __snake_case , __snake_case : str = jnp.split(lowerCamelCase , sample.shape[1] , axis=1 ) else: __snake_case : List[Any] = None # 1. compute alphas, betas __snake_case : Optional[Any] = state.common.alphas_cumprod[t] __snake_case : List[Any] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) __snake_case : List[Any] = 1 - alpha_prod_t __snake_case : Union[str, Any] = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": __snake_case : int = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": __snake_case : Optional[Any] = model_output elif self.config.prediction_type == "v_prediction": __snake_case : List[str] = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F'prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ' " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: __snake_case : List[str] = jnp.clip(lowerCamelCase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __snake_case : Tuple = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t __snake_case : str = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf __snake_case : int = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): __snake_case : List[Any] = jax.random.split(lowerCamelCase , num=1 ) __snake_case : Dict = jax.random.normal(lowerCamelCase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(lowerCamelCase , lowerCamelCase , predicted_variance=lowerCamelCase ) ** 0.5) * noise __snake_case : Union[str, Any] = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) __snake_case : int = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=lowerCamelCase , state=lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : DDPMSchedulerState , lowerCamelCase : jnp.ndarray , lowerCamelCase : jnp.ndarray , lowerCamelCase : jnp.ndarray , ) -> jnp.ndarray: return add_noise_common(state.common , lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : DDPMSchedulerState , lowerCamelCase : jnp.ndarray , lowerCamelCase : jnp.ndarray , lowerCamelCase : jnp.ndarray , ) -> jnp.ndarray: return get_velocity_common(state.common , lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __len__( self : Dict ) -> Dict: return self.config.num_train_timesteps
81
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
1
import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING _snake_case : List[str] = logging.get_logger(__name__) class a (enum.Enum ): """simple docstring""" __UpperCAmelCase : str = 0 __UpperCAmelCase : str = 1 @add_end_docstrings(_lowerCAmelCase ) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Any = "generated" def __init__( self : List[Any] , *lowerCamelCase : List[str] , **lowerCamelCase : str ) -> Any: super().__init__(*lowerCamelCase , **lowerCamelCase ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def __snake_case ( self : Any , lowerCamelCase : Union[str, Any]=None , lowerCamelCase : Union[str, Any]=None , lowerCamelCase : Union[str, Any]=None , lowerCamelCase : List[Any]=None , lowerCamelCase : Union[str, Any]=None , lowerCamelCase : Tuple=None , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : List[Any] = {} if truncation is not None: __snake_case : Optional[int] = truncation __snake_case : int = generate_kwargs __snake_case : str = {} if return_tensors is not None and return_type is None: __snake_case : Dict = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: __snake_case : Optional[int] = return_type if clean_up_tokenization_spaces is not None: __snake_case : Any = clean_up_tokenization_spaces if stop_sequence is not None: __snake_case : Optional[Any] = self.tokenizer.encode(lowerCamelCase , add_special_tokens=lowerCamelCase ) if len(lowerCamelCase ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) __snake_case : str = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __snake_case ( self : Union[str, Any] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> List[Any]: return True def __snake_case ( self : Any , *lowerCamelCase : Dict , lowerCamelCase : Any ) -> int: __snake_case : Union[str, Any] = self.model.config.prefix if self.model.config.prefix is not None else "" if isinstance(args[0] , lowerCamelCase ): if self.tokenizer.pad_token_id is None: raise ValueError("Please make sure that the tokenizer has a pad_token_id when using a batch input" ) __snake_case : str = ([prefix + arg for arg in args[0]],) __snake_case : List[str] = True elif isinstance(args[0] , lowerCamelCase ): __snake_case : List[str] = (prefix + args[0],) __snake_case : Optional[int] = False else: raise ValueError( F' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`' ) __snake_case : str = self.tokenizer(*lowerCamelCase , padding=lowerCamelCase , truncation=lowerCamelCase , return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : Optional[int] , *lowerCamelCase : Tuple , **lowerCamelCase : str ) -> List[Any]: __snake_case : Any = super().__call__(*lowerCamelCase , **lowerCamelCase ) if ( isinstance(args[0] , lowerCamelCase ) and all(isinstance(lowerCamelCase , lowerCamelCase ) for el in args[0] ) and all(len(lowerCamelCase ) == 1 for res in result ) ): return [res[0] for res in result] return result def __snake_case ( self : int , lowerCamelCase : Union[str, Any] , lowerCamelCase : List[str]=TruncationStrategy.DO_NOT_TRUNCATE , **lowerCamelCase : str ) -> int: __snake_case : int = self._parse_and_tokenize(lowerCamelCase , truncation=lowerCamelCase , **lowerCamelCase ) return inputs def __snake_case ( self : Optional[Any] , lowerCamelCase : Optional[Any] , **lowerCamelCase : str ) -> Any: if self.framework == "pt": __snake_case , __snake_case : Dict = model_inputs["input_ids"].shape elif self.framework == "tf": __snake_case , __snake_case : Optional[Any] = tf.shape(model_inputs["input_ids"] ).numpy() __snake_case : Union[str, Any] = generate_kwargs.get("min_length" , self.model.config.min_length ) __snake_case : int = generate_kwargs.get("max_length" , self.model.config.max_length ) self.check_inputs(lowerCamelCase , generate_kwargs["min_length"] , generate_kwargs["max_length"] ) __snake_case : Tuple = self.model.generate(**lowerCamelCase , **lowerCamelCase ) __snake_case : Union[str, Any] = output_ids.shape[0] if self.framework == "pt": __snake_case : int = output_ids.reshape(lowerCamelCase , out_b // in_b , *output_ids.shape[1:] ) elif self.framework == "tf": __snake_case : Tuple = tf.reshape(lowerCamelCase , (in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def __snake_case ( self : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any]=ReturnType.TEXT , lowerCamelCase : List[str]=False ) -> List[str]: __snake_case : List[Any] = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: __snake_case : str = {F'{self.return_name}_token_ids': output_ids} elif return_type == ReturnType.TEXT: __snake_case : List[Any] = { F'{self.return_name}_text': self.tokenizer.decode( lowerCamelCase , skip_special_tokens=lowerCamelCase , clean_up_tokenization_spaces=lowerCamelCase , ) } records.append(lowerCamelCase ) return records @add_end_docstrings(_lowerCAmelCase ) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : int = "summary" def __call__( self : Dict , *lowerCamelCase : Tuple , **lowerCamelCase : Optional[int] ) -> Any: return super().__call__(*lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> bool: if max_length < min_length: logger.warning(F'Your min_length={min_length} must be inferior than your max_length={max_length}.' ) if input_length < max_length: logger.warning( F'Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ' "a summarization task, where outputs shorter than the input are typically wanted, you might " F'consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})' ) @add_end_docstrings(_lowerCAmelCase ) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = "translation" def __snake_case ( self : List[Any] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> int: if input_length > 0.9 * max_length: logger.warning( F'Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ' "increasing your max_length manually, e.g. translator('...', max_length=400)" ) return True def __snake_case ( self : Any , *lowerCamelCase : List[str] , lowerCamelCase : Dict=TruncationStrategy.DO_NOT_TRUNCATE , lowerCamelCase : Optional[int]=None , lowerCamelCase : Optional[Any]=None ) -> List[str]: if getattr(self.tokenizer , "_build_translation_inputs" , lowerCamelCase ): return self.tokenizer._build_translation_inputs( *lowerCamelCase , return_tensors=self.framework , truncation=lowerCamelCase , src_lang=lowerCamelCase , tgt_lang=lowerCamelCase ) else: return super()._parse_and_tokenize(*lowerCamelCase , truncation=lowerCamelCase ) def __snake_case ( self : str , lowerCamelCase : str=None , lowerCamelCase : Dict=None , **lowerCamelCase : List[Any] ) -> Optional[int]: __snake_case , __snake_case , __snake_case : int = super()._sanitize_parameters(**lowerCamelCase ) if src_lang is not None: __snake_case : Optional[Any] = src_lang if tgt_lang is not None: __snake_case : str = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. __snake_case : Union[str, Any] = kwargs.get("task" , self.task ) __snake_case : str = task.split("_" ) if task and len(lowerCamelCase ) == 4: # translation, XX, to YY __snake_case : Optional[int] = items[1] __snake_case : Dict = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : Tuple , *lowerCamelCase : int , **lowerCamelCase : List[str] ) -> str: return super().__call__(*lowerCamelCase , **lowerCamelCase )
81
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
1
import argparse import logging import pickle from collections import Counter logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO ) _snake_case : Dict = logging.getLogger(__name__) if __name__ == "__main__": _snake_case : int = argparse.ArgumentParser( description="Token Counts for smoothing the masking probabilities in MLM (cf XLM/word2vec)" ) parser.add_argument( "--data_file", type=str, default="data/dump.bert-base-uncased.pickle", help="The binarized dataset." ) parser.add_argument( "--token_counts_dump", type=str, default="data/token_counts.bert-base-uncased.pickle", help="The dump file." ) parser.add_argument("--vocab_size", default=30_522, type=int) _snake_case : str = parser.parse_args() logger.info(f'''Loading data from {args.data_file}''') with open(args.data_file, "rb") as fp: _snake_case : str = pickle.load(fp) logger.info("Counting occurrences for MLM.") _snake_case : Optional[int] = Counter() for tk_ids in data: counter.update(tk_ids) _snake_case : int = [0] * args.vocab_size for k, v in counter.items(): _snake_case : List[Any] = v logger.info(f'''Dump to {args.token_counts_dump}''') with open(args.token_counts_dump, "wb") as handle: pickle.dump(counts, handle, protocol=pickle.HIGHEST_PROTOCOL)
81
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
1
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : torch.FloatTensor class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : List[Any]=3 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=("DownEncoderBlock2D",) , lowerCamelCase : str=(64,) , lowerCamelCase : Optional[Any]=2 , lowerCamelCase : Any=32 , lowerCamelCase : Any="silu" , lowerCamelCase : List[str]=True , ) -> str: super().__init__() __snake_case : Union[str, Any] = layers_per_block __snake_case : str = torch.nn.Convad( lowerCamelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) __snake_case : Optional[int] = None __snake_case : Any = nn.ModuleList([] ) # down __snake_case : Any = block_out_channels[0] for i, down_block_type in enumerate(lowerCamelCase ): __snake_case : List[str] = output_channel __snake_case : Any = block_out_channels[i] __snake_case : Tuple = i == len(lowerCamelCase ) - 1 __snake_case : Dict = get_down_block( lowerCamelCase , num_layers=self.layers_per_block , in_channels=lowerCamelCase , out_channels=lowerCamelCase , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCamelCase , resnet_groups=lowerCamelCase , attention_head_dim=lowerCamelCase , temb_channels=lowerCamelCase , ) self.down_blocks.append(lowerCamelCase ) # mid __snake_case : str = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCamelCase , output_scale_factor=1 , resnet_time_scale_shift="default" , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCamelCase , temb_channels=lowerCamelCase , ) # out __snake_case : Optional[int] = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCamelCase , eps=1E-6 ) __snake_case : int = nn.SiLU() __snake_case : Tuple = 2 * out_channels if double_z else out_channels __snake_case : str = nn.Convad(block_out_channels[-1] , lowerCamelCase , 3 , padding=1 ) __snake_case : Any = False def __snake_case ( self : List[Any] , lowerCamelCase : int ) -> int: __snake_case : Any = x __snake_case : Union[str, Any] = self.conv_in(lowerCamelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCamelCase : Dict ): def custom_forward(*lowerCamelCase : str ): return module(*lowerCamelCase ) return custom_forward # down if is_torch_version(">=" , "1.11.0" ): for down_block in self.down_blocks: __snake_case : str = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCamelCase ) , lowerCamelCase , use_reentrant=lowerCamelCase ) # middle __snake_case : Optional[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase , use_reentrant=lowerCamelCase ) else: for down_block in self.down_blocks: __snake_case : Union[str, Any] = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCamelCase ) , lowerCamelCase ) # middle __snake_case : List[str] = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCamelCase ) else: # down for down_block in self.down_blocks: __snake_case : Union[str, Any] = down_block(lowerCamelCase ) # middle __snake_case : Dict = self.mid_block(lowerCamelCase ) # post-process __snake_case : Union[str, Any] = self.conv_norm_out(lowerCamelCase ) __snake_case : int = self.conv_act(lowerCamelCase ) __snake_case : List[Any] = self.conv_out(lowerCamelCase ) return sample class a (nn.Module ): """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : int=3 , lowerCamelCase : int=3 , lowerCamelCase : Any=("UpDecoderBlock2D",) , lowerCamelCase : List[str]=(64,) , lowerCamelCase : List[str]=2 , lowerCamelCase : Any=32 , lowerCamelCase : Union[str, Any]="silu" , lowerCamelCase : Union[str, Any]="group" , ) -> str: super().__init__() __snake_case : Tuple = layers_per_block __snake_case : Optional[int] = nn.Convad( lowerCamelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) __snake_case : Dict = None __snake_case : Optional[int] = nn.ModuleList([] ) __snake_case : List[Any] = in_channels if norm_type == "spatial" else None # mid __snake_case : int = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCamelCase , output_scale_factor=1 , resnet_time_scale_shift="default" if norm_type == "group" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCamelCase , temb_channels=lowerCamelCase , ) # up __snake_case : Dict = list(reversed(lowerCamelCase ) ) __snake_case : List[str] = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCamelCase ): __snake_case : Any = output_channel __snake_case : Tuple = reversed_block_out_channels[i] __snake_case : int = i == len(lowerCamelCase ) - 1 __snake_case : List[str] = get_up_block( lowerCamelCase , num_layers=self.layers_per_block + 1 , in_channels=lowerCamelCase , out_channels=lowerCamelCase , prev_output_channel=lowerCamelCase , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCamelCase , resnet_groups=lowerCamelCase , attention_head_dim=lowerCamelCase , temb_channels=lowerCamelCase , resnet_time_scale_shift=lowerCamelCase , ) self.up_blocks.append(lowerCamelCase ) __snake_case : int = output_channel # out if norm_type == "spatial": __snake_case : List[Any] = SpatialNorm(block_out_channels[0] , lowerCamelCase ) else: __snake_case : List[str] = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCamelCase , eps=1E-6 ) __snake_case : Any = nn.SiLU() __snake_case : Optional[Any] = nn.Convad(block_out_channels[0] , lowerCamelCase , 3 , padding=1 ) __snake_case : Any = False def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Dict=None ) -> int: __snake_case : Union[str, Any] = z __snake_case : int = self.conv_in(lowerCamelCase ) __snake_case : Optional[int] = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCamelCase : str ): def custom_forward(*lowerCamelCase : Optional[Any] ): return module(*lowerCamelCase ) return custom_forward if is_torch_version(">=" , "1.11.0" ): # middle __snake_case : str = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase , lowerCamelCase , use_reentrant=lowerCamelCase ) __snake_case : Dict = sample.to(lowerCamelCase ) # up for up_block in self.up_blocks: __snake_case : str = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCamelCase ) , lowerCamelCase , lowerCamelCase , use_reentrant=lowerCamelCase ) else: # middle __snake_case : Union[str, Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCamelCase , lowerCamelCase ) __snake_case : List[Any] = sample.to(lowerCamelCase ) # up for up_block in self.up_blocks: __snake_case : Any = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCamelCase ) , lowerCamelCase , lowerCamelCase ) else: # middle __snake_case : Tuple = self.mid_block(lowerCamelCase , lowerCamelCase ) __snake_case : Union[str, Any] = sample.to(lowerCamelCase ) # up for up_block in self.up_blocks: __snake_case : int = up_block(lowerCamelCase , lowerCamelCase ) # post-process if latent_embeds is None: __snake_case : List[str] = self.conv_norm_out(lowerCamelCase ) else: __snake_case : List[Any] = self.conv_norm_out(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = self.conv_act(lowerCamelCase ) __snake_case : List[Any] = self.conv_out(lowerCamelCase ) return sample class a (nn.Module ): """simple docstring""" def __init__( self : Optional[Any] , lowerCamelCase : int , lowerCamelCase : Union[str, Any] , lowerCamelCase : List[Any] , lowerCamelCase : Tuple=None , lowerCamelCase : Dict="random" , lowerCamelCase : List[str]=False , lowerCamelCase : int=True ) -> List[str]: super().__init__() __snake_case : Any = n_e __snake_case : Optional[int] = vq_embed_dim __snake_case : List[str] = beta __snake_case : List[Any] = legacy __snake_case : Tuple = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) __snake_case : Union[str, Any] = remap if self.remap is not None: self.register_buffer("used" , torch.tensor(np.load(self.remap ) ) ) __snake_case : List[str] = self.used.shape[0] __snake_case : Union[str, Any] = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": __snake_case : List[str] = self.re_embed __snake_case : str = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: __snake_case : List[Any] = n_e __snake_case : Union[str, Any] = sane_index_shape def __snake_case ( self : str , lowerCamelCase : Optional[Any] ) -> int: __snake_case : Union[str, Any] = inds.shape assert len(lowerCamelCase ) > 1 __snake_case : Tuple = inds.reshape(ishape[0] , -1 ) __snake_case : Union[str, Any] = self.used.to(lowerCamelCase ) __snake_case : Union[str, Any] = (inds[:, :, None] == used[None, None, ...]).long() __snake_case : Optional[Any] = match.argmax(-1 ) __snake_case : Tuple = match.sum(2 ) < 1 if self.unknown_index == "random": __snake_case : List[str] = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: __snake_case : Any = self.unknown_index return new.reshape(lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : int ) -> List[Any]: __snake_case : str = inds.shape assert len(lowerCamelCase ) > 1 __snake_case : Union[str, Any] = inds.reshape(ishape[0] , -1 ) __snake_case : Any = self.used.to(lowerCamelCase ) if self.re_embed > self.used.shape[0]: # extra token __snake_case : List[str] = 0 # simply set to zero __snake_case : List[Any] = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCamelCase ) return back.reshape(lowerCamelCase ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Tuple ) -> Any: # reshape z -> (batch, height, width, channel) and flatten __snake_case : Any = z.permute(0 , 2 , 3 , 1 ).contiguous() __snake_case : Optional[Any] = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z __snake_case : Optional[int] = torch.argmin(torch.cdist(lowerCamelCase , self.embedding.weight ) , dim=1 ) __snake_case : Union[str, Any] = self.embedding(lowerCamelCase ).view(z.shape ) __snake_case : Any = None __snake_case : Optional[Any] = None # compute loss for embedding if not self.legacy: __snake_case : Union[str, Any] = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: __snake_case : int = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients __snake_case : Dict = z + (z_q - z).detach() # reshape back to match original input shape __snake_case : Dict = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: __snake_case : int = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis __snake_case : Union[str, Any] = self.remap_to_used(lowerCamelCase ) __snake_case : Union[str, Any] = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: __snake_case : List[Any] = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def __snake_case ( self : Optional[Any] , lowerCamelCase : Dict , lowerCamelCase : Dict ) -> List[str]: # shape specifying (batch, height, width, channel) if self.remap is not None: __snake_case : List[Any] = indices.reshape(shape[0] , -1 ) # add batch axis __snake_case : Dict = self.unmap_to_all(lowerCamelCase ) __snake_case : List[Any] = indices.reshape(-1 ) # flatten again # get quantized latent vectors __snake_case : Optional[Any] = self.embedding(lowerCamelCase ) if shape is not None: __snake_case : List[Any] = z_q.view(lowerCamelCase ) # reshape back to match original input shape __snake_case : Optional[int] = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : List[Any]=False ) -> Optional[Any]: __snake_case : List[str] = parameters __snake_case , __snake_case : List[Any] = torch.chunk(lowerCamelCase , 2 , dim=1 ) __snake_case : int = torch.clamp(self.logvar , -30.0 , 20.0 ) __snake_case : List[Any] = deterministic __snake_case : Tuple = torch.exp(0.5 * self.logvar ) __snake_case : Tuple = torch.exp(self.logvar ) if self.deterministic: __snake_case : str = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def __snake_case ( self : int , lowerCamelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: # make sure sample is on the same device as the parameters and has same dtype __snake_case : Tuple = randn_tensor( self.mean.shape , generator=lowerCamelCase , device=self.parameters.device , dtype=self.parameters.dtype ) __snake_case : Optional[int] = self.mean + self.std * sample return x def __snake_case ( self : List[Any] , lowerCamelCase : Union[str, Any]=None ) -> Dict: if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def __snake_case ( self : str , lowerCamelCase : Union[str, Any] , lowerCamelCase : Optional[Any]=[1, 2, 3] ) -> List[str]: if self.deterministic: return torch.Tensor([0.0] ) __snake_case : List[str] = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCamelCase ) def __snake_case ( self : int ) -> Union[str, Any]: return self.mean
81
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
1
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : Dict=3 , lowerCamelCase : Optional[Any]=18 , lowerCamelCase : Optional[Any]=30 , lowerCamelCase : Any=400 , lowerCamelCase : int=True , lowerCamelCase : Tuple=None , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=None , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : str=[0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , lowerCamelCase : Optional[int]=[0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , lowerCamelCase : Optional[int]=True , ) -> List[Any]: __snake_case : List[str] = size if size is not None else {"height": 224, "width": 224} __snake_case : List[str] = crop_size if crop_size is not None else {"height": 18, "width": 18} __snake_case : Tuple = parent __snake_case : List[str] = batch_size __snake_case : Union[str, Any] = num_channels __snake_case : Union[str, Any] = image_size __snake_case : Union[str, Any] = min_resolution __snake_case : str = max_resolution __snake_case : List[Any] = do_resize __snake_case : Optional[int] = size __snake_case : int = do_center_crop __snake_case : Dict = crop_size __snake_case : List[Any] = do_normalize __snake_case : str = image_mean __snake_case : Optional[int] = image_std __snake_case : Union[str, Any] = do_convert_rgb def __snake_case ( self : str ) -> Optional[int]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def __snake_case ( self : Optional[Any] , lowerCamelCase : List[Any]=False , lowerCamelCase : Optional[Any]=False , lowerCamelCase : List[str]=False ) -> Optional[int]: assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: __snake_case : List[str] = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 255 , size=(self.num_channels, self.max_resolution, self.max_resolution) , dtype=np.uinta ) ) else: __snake_case : List[str] = [] for i in range(self.batch_size ): __snake_case , __snake_case : Tuple = np.random.choice(np.arange(self.min_resolution , self.max_resolution ) , 2 ) image_inputs.append(np.random.randint(255 , size=(self.num_channels, width, height) , dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension __snake_case : str = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] if torchify: __snake_case : Dict = [torch.from_numpy(lowerCamelCase ) for x in image_inputs] return image_inputs @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Tuple = ChineseCLIPImageProcessor if is_vision_available() else None def __snake_case ( self : int ) -> str: __snake_case : Optional[Any] = ChineseCLIPImageProcessingTester(self , do_center_crop=lowerCamelCase ) @property def __snake_case ( self : List[str] ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Union[str, Any] ) -> Optional[Any]: __snake_case : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_convert_rgb" ) ) def __snake_case ( self : List[Any] ) -> Dict: __snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 224, "width": 224} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) __snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __snake_case ( self : str ) -> int: pass def __snake_case ( self : Optional[int] ) -> Any: # Initialize image_processing __snake_case : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : List[Any] = self.image_processor_tester.prepare_inputs(equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : Optional[int] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __snake_case ( self : Optional[Any] ) -> Optional[int]: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = self.image_processor_tester.prepare_inputs(equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Any = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched __snake_case : List[str] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __snake_case ( self : Any ) -> Dict: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : str = self.image_processor_tester.prepare_inputs(equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = ChineseCLIPImageProcessor if is_vision_available() else None def __snake_case ( self : Tuple ) -> Optional[int]: __snake_case : Optional[int] = ChineseCLIPImageProcessingTester(self , num_channels=4 , do_center_crop=lowerCamelCase ) __snake_case : Any = 3 @property def __snake_case ( self : List[Any] ) -> Dict: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : List[str] ) -> Union[str, Any]: __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_convert_rgb" ) ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : List[Any] ) -> Any: # Initialize image_processing __snake_case : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : Union[str, Any] = self.image_processor_tester.prepare_inputs(equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
81
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if len(__lowerCamelCase ) == 0: return array __snake_case , __snake_case : List[Any] = min(__lowerCamelCase ), max(__lowerCamelCase ) # Compute the variables __snake_case : int = _max - _min + 1 __snake_case , __snake_case : Tuple = [0] * holes_range, [0] * holes_range # Make the sorting. for i in array: __snake_case : Any = i - _min __snake_case : int = i holes_repeat[index] += 1 # Makes the array back by replacing the numbers. __snake_case : List[Any] = 0 for i in range(__lowerCamelCase ): while holes_repeat[i] > 0: __snake_case : List[str] = holes[i] index += 1 holes_repeat[i] -= 1 # Returns the sorted array. return array if __name__ == "__main__": import doctest doctest.testmod() _snake_case : Optional[Any] = input("Enter numbers separated by comma:\n") _snake_case : List[str] = [int(x) for x in user_input.split(",")] print(pigeon_sort(unsorted))
81
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
1
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class a : """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , ) -> Tuple: __snake_case : Union[str, Any] = parent __snake_case : Dict = 13 __snake_case : Tuple = 7 __snake_case : Dict = 30 __snake_case : Any = self.seq_length + self.mem_len __snake_case : int = 15 __snake_case : Union[str, Any] = True __snake_case : List[Any] = True __snake_case : List[Any] = 99 __snake_case : Optional[int] = [10, 50, 80] __snake_case : List[str] = 32 __snake_case : List[Any] = 32 __snake_case : int = 4 __snake_case : Optional[Any] = 8 __snake_case : Union[str, Any] = 128 __snake_case : Optional[Any] = 2 __snake_case : Any = 2 __snake_case : str = None __snake_case : Optional[int] = 1 __snake_case : str = 0 __snake_case : Union[str, Any] = 3 __snake_case : Tuple = self.vocab_size - 1 __snake_case : List[str] = 0.01 def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : int = None if self.use_labels: __snake_case : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : Dict = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def __snake_case ( self : Any ) -> Union[str, Any]: random.seed(self.seed ) tf.random.set_seed(self.seed ) def __snake_case ( self : str , lowerCamelCase : Tuple , lowerCamelCase : Union[str, Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : str ) -> Optional[Any]: __snake_case : List[str] = TFTransfoXLModel(lowerCamelCase ) __snake_case , __snake_case : str = model(lowerCamelCase ).to_tuple() __snake_case : Optional[Any] = {"input_ids": input_ids_a, "mems": mems_a} __snake_case , __snake_case : List[Any] = model(lowerCamelCase ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[int] , lowerCamelCase : Tuple , lowerCamelCase : List[Any] , lowerCamelCase : int ) -> Dict: __snake_case : int = TFTransfoXLLMHeadModel(lowerCamelCase ) __snake_case , __snake_case : str = model(lowerCamelCase ).to_tuple() __snake_case : Tuple = {"input_ids": input_ids_a, "labels": lm_labels} __snake_case , __snake_case : List[Any] = model(lowerCamelCase ).to_tuple() __snake_case , __snake_case : Union[str, Any] = model([input_ids_a, mems_a] ).to_tuple() __snake_case : List[str] = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} __snake_case , __snake_case : List[Any] = model(lowerCamelCase ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : str , lowerCamelCase : Optional[int] , lowerCamelCase : str , lowerCamelCase : Optional[int] ) -> List[str]: __snake_case : str = TFTransfoXLForSequenceClassification(lowerCamelCase ) __snake_case : Optional[int] = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : List[Any] ) -> Any: __snake_case : Union[str, Any] = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) : Tuple = config_and_inputs __snake_case : Dict = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Any = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) __UpperCAmelCase : List[Any] = () if is_tf_available() else () __UpperCAmelCase : Dict = ( { "feature-extraction": TFTransfoXLModel, "text-classification": TFTransfoXLForSequenceClassification, "text-generation": TFTransfoXLLMHeadModel, "zero-shot": TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : List[Any] = False __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : int = False def __snake_case ( self : Optional[int] , lowerCamelCase : int , lowerCamelCase : List[str] , lowerCamelCase : List[str] , lowerCamelCase : str , lowerCamelCase : Any ) -> Optional[int]: if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def __snake_case ( self : int ) -> int: __snake_case : List[Any] = TFTransfoXLModelTester(self ) __snake_case : List[str] = ConfigTester(self , config_class=lowerCamelCase , d_embed=37 ) def __snake_case ( self : int ) -> List[Any]: self.config_tester.run_common_tests() def __snake_case ( self : List[Any] ) -> Dict: self.model_tester.set_seed() __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Optional[Any]: self.model_tester.set_seed() __snake_case : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Tuple: __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : int = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: __snake_case : str = model_class(lowerCamelCase ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: __snake_case : int = model.get_output_embeddings() assert isinstance(lowerCamelCase , tf.keras.layers.Layer ) __snake_case : Optional[Any] = model.get_bias() assert name is None else: __snake_case : Optional[int] = model.get_output_embeddings() assert x is None __snake_case : Tuple = model.get_bias() assert name is None def __snake_case ( self : List[Any] ) -> List[Any]: # TODO JP: Make TransfoXL XLA compliant pass @slow def __snake_case ( self : Optional[int] ) -> Dict: for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Union[str, Any] = TFTransfoXLModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @unittest.skip(reason="This model doesn't play well with fit() due to not returning a single loss." ) def __snake_case ( self : str ) -> Optional[int]: pass @require_tf class a (unittest.TestCase ): """simple docstring""" @unittest.skip("Skip test until #12651 is resolved." ) @slow def __snake_case ( self : List[Any] ) -> Tuple: __snake_case : Dict = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103" ) # fmt: off __snake_case : Union[str, Any] = tf.convert_to_tensor([[33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off __snake_case : List[str] = [33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0,33,1,1857,2,1,1009,4,1109,11739,4762,358,5,25,245,28,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> __snake_case : Optional[int] = model.generate(lowerCamelCase , max_length=200 , do_sample=lowerCamelCase ) self.assertListEqual(output_ids[0].numpy().tolist() , lowerCamelCase )
81
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _snake_case : Optional[Any] = logging.get_logger(__name__) _snake_case : Optional[int] = { "xlm-mlm-en-2048": "https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json", "xlm-mlm-ende-1024": "https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json", "xlm-mlm-enfr-1024": "https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json", "xlm-mlm-enro-1024": "https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json", "xlm-mlm-tlm-xnli15-1024": "https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json", "xlm-mlm-xnli15-1024": "https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json", "xlm-clm-enfr-1024": "https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json", "xlm-clm-ende-1024": "https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json", "xlm-mlm-17-1280": "https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json", "xlm-mlm-100-1280": "https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json", } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Any = "xlm" __UpperCAmelCase : List[str] = { "hidden_size": "emb_dim", "num_attention_heads": "n_heads", "num_hidden_layers": "n_layers", "n_words": "vocab_size", # For backward compatibility } def __init__( self : List[str] , lowerCamelCase : Union[str, Any]=30145 , lowerCamelCase : Union[str, Any]=2048 , lowerCamelCase : Any=12 , lowerCamelCase : Any=16 , lowerCamelCase : List[str]=0.1 , lowerCamelCase : List[str]=0.1 , lowerCamelCase : Optional[int]=True , lowerCamelCase : str=False , lowerCamelCase : Optional[Any]=False , lowerCamelCase : int=False , lowerCamelCase : Optional[Any]=1 , lowerCamelCase : Optional[int]=True , lowerCamelCase : Dict=512 , lowerCamelCase : Tuple=2048**-0.5 , lowerCamelCase : Optional[int]=1E-12 , lowerCamelCase : int=0.02 , lowerCamelCase : Union[str, Any]=0 , lowerCamelCase : List[str]=1 , lowerCamelCase : Dict=2 , lowerCamelCase : Optional[Any]=3 , lowerCamelCase : Union[str, Any]=5 , lowerCamelCase : Dict=True , lowerCamelCase : List[Any]="first" , lowerCamelCase : Dict=True , lowerCamelCase : Any=None , lowerCamelCase : List[str]=True , lowerCamelCase : List[str]=0.1 , lowerCamelCase : Optional[Any]=5 , lowerCamelCase : int=5 , lowerCamelCase : int=0 , lowerCamelCase : Optional[int]=0 , lowerCamelCase : Union[str, Any]=2 , lowerCamelCase : List[str]=0 , **lowerCamelCase : Optional[int] , ) -> List[Any]: __snake_case : Any = vocab_size __snake_case : List[Any] = emb_dim __snake_case : List[Any] = n_layers __snake_case : Dict = n_heads __snake_case : Any = dropout __snake_case : List[str] = attention_dropout __snake_case : Any = gelu_activation __snake_case : Dict = sinusoidal_embeddings __snake_case : Optional[int] = causal __snake_case : Dict = asm __snake_case : Optional[int] = n_langs __snake_case : Dict = use_lang_emb __snake_case : str = layer_norm_eps __snake_case : Optional[Any] = bos_index __snake_case : Optional[Any] = eos_index __snake_case : List[Any] = pad_index __snake_case : str = unk_index __snake_case : int = mask_index __snake_case : Tuple = is_encoder __snake_case : Optional[int] = max_position_embeddings __snake_case : Any = embed_init_std __snake_case : Optional[int] = init_std __snake_case : Optional[int] = summary_type __snake_case : Optional[int] = summary_use_proj __snake_case : Tuple = summary_activation __snake_case : str = summary_proj_to_labels __snake_case : Union[str, Any] = summary_first_dropout __snake_case : Union[str, Any] = start_n_top __snake_case : List[Any] = end_n_top __snake_case : Dict = mask_token_id __snake_case : Tuple = lang_id if "n_words" in kwargs: __snake_case : str = kwargs["n_words"] super().__init__(pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , **lowerCamelCase ) class a (_lowerCAmelCase ): """simple docstring""" @property def __snake_case ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": __snake_case : Optional[int] = {0: "batch", 1: "choice", 2: "sequence"} else: __snake_case : List[str] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
81
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
1
import math def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = len(__lowerCamelCase ) __snake_case : Tuple = int(math.floor(math.sqrt(__lowerCamelCase ) ) ) __snake_case : Optional[Any] = 0 while arr[min(__lowerCamelCase , __lowerCamelCase ) - 1] < x: __snake_case : Optional[int] = step step += int(math.floor(math.sqrt(__lowerCamelCase ) ) ) if prev >= n: return -1 while arr[prev] < x: __snake_case : Optional[int] = prev + 1 if prev == min(__lowerCamelCase , __lowerCamelCase ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": _snake_case : int = input("Enter numbers separated by a comma:\n").strip() _snake_case : Tuple = [int(item) for item in user_input.split(",")] _snake_case : str = int(input("Enter the number to be searched:\n")) _snake_case : List[str] = jump_search(arr, x) if res == -1: print("Number not found!") else: print(f'''Number {x} is at index {res}''')
81
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
1
import string def lowerCAmelCase_ ( __lowerCamelCase ): for key in range(len(string.ascii_uppercase ) ): __snake_case : Optional[int] = "" for symbol in message: if symbol in string.ascii_uppercase: __snake_case : Dict = string.ascii_uppercase.find(__lowerCamelCase ) __snake_case : Any = num - key if num < 0: __snake_case : Optional[Any] = num + len(string.ascii_uppercase ) __snake_case : List[Any] = translated + string.ascii_uppercase[num] else: __snake_case : Dict = translated + symbol print(F'Decryption using Key #{key}: {translated}' ) def lowerCAmelCase_ ( ): __snake_case : Tuple = input("Encrypted message: " ) __snake_case : int = message.upper() decrypt(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
81
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
1
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(sqrt(__lowerCamelCase ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_1 ): __snake_case : List[Any] = 0 __snake_case : Union[str, Any] = 1 while count != nth and number < 3: number += 1 if is_prime(__lowerCamelCase ): count += 1 while count != nth: number += 2 if is_prime(__lowerCamelCase ): count += 1 return number if __name__ == "__main__": print(f'''{solution() = }''')
81
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
1
import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> List[str]: __snake_case : List[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Any , lowerCamelCase : Dict , lowerCamelCase : List[str]=13 , lowerCamelCase : Union[str, Any]=64 , lowerCamelCase : Any=3 , lowerCamelCase : int=3 , lowerCamelCase : Union[str, Any]=2 , lowerCamelCase : Tuple=1 , lowerCamelCase : Optional[Any]=16 , lowerCamelCase : Dict=[128, 256, 384] , lowerCamelCase : Union[str, Any]=[4, 6, 8] , lowerCamelCase : Any=[2, 3, 4] , lowerCamelCase : int=[16, 16, 16] , lowerCamelCase : List[str]=0 , lowerCamelCase : List[Any]=[2, 2, 2] , lowerCamelCase : List[str]=[2, 2, 2] , lowerCamelCase : Tuple=0.02 , lowerCamelCase : int=True , lowerCamelCase : Any=True , lowerCamelCase : Any=2 , ) -> Optional[int]: __snake_case : Any = parent __snake_case : List[Any] = batch_size __snake_case : Any = image_size __snake_case : Optional[int] = num_channels __snake_case : Union[str, Any] = kernel_size __snake_case : Dict = stride __snake_case : Any = padding __snake_case : Tuple = hidden_sizes __snake_case : Optional[Any] = num_attention_heads __snake_case : int = depths __snake_case : int = key_dim __snake_case : Any = drop_path_rate __snake_case : List[Any] = patch_size __snake_case : str = attention_ratio __snake_case : Optional[Any] = mlp_ratio __snake_case : str = initializer_range __snake_case : Dict = [ ["Subsample", key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ["Subsample", key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] __snake_case : str = is_training __snake_case : Any = use_labels __snake_case : List[str] = num_labels __snake_case : Any = initializer_range def __snake_case ( self : Tuple ) -> Optional[int]: __snake_case : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Any = None if self.use_labels: __snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : int = self.get_config() return config, pixel_values, labels def __snake_case ( self : Dict ) -> Optional[int]: return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : Optional[Any] , lowerCamelCase : str ) -> str: __snake_case : List[Any] = LevitModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Optional[Any] = model(lowerCamelCase ) __snake_case : Dict = (self.image_size, self.image_size) __snake_case , __snake_case : List[str] = image_size[0], image_size[1] for _ in range(4 ): __snake_case : int = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) __snake_case : Optional[int] = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) , ) def __snake_case ( self : List[str] , lowerCamelCase : str , lowerCamelCase : Any , lowerCamelCase : int ) -> Optional[int]: __snake_case : int = self.num_labels __snake_case : int = LevitForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : str = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Any ) -> Dict: __snake_case : Any = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case : str = config_and_inputs __snake_case : Any = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) __UpperCAmelCase : Optional[int] = ( { "feature-extraction": LevitModel, "image-classification": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) __UpperCAmelCase : Dict = False __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : int = False __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : Tuple = False def __snake_case ( self : Optional[Any] ) -> Any: __snake_case : int = LevitModelTester(self ) __snake_case : Optional[int] = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def __snake_case ( self : Optional[int] ) -> Tuple: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __snake_case ( self : Dict ) -> str: return @unittest.skip(reason="Levit does not use inputs_embeds" ) def __snake_case ( self : Any ) -> Optional[int]: pass @unittest.skip(reason="Levit does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> int: pass @unittest.skip(reason="Levit does not output attentions" ) def __snake_case ( self : Optional[int] ) -> Optional[int]: pass def __snake_case ( self : Tuple ) -> Tuple: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Tuple = [*signature.parameters.keys()] __snake_case : Any = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : int ) -> Tuple: def check_hidden_states_output(lowerCamelCase : str , lowerCamelCase : Tuple , lowerCamelCase : Any ): __snake_case : List[str] = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : Optional[Any] = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : str = outputs.hidden_states __snake_case : Tuple = len(self.model_tester.depths ) + 1 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) __snake_case : Optional[int] = (self.model_tester.image_size, self.model_tester.image_size) __snake_case , __snake_case : List[str] = image_size[0], image_size[1] for _ in range(4 ): __snake_case : Tuple = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) __snake_case : Dict = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) __snake_case , __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : Tuple = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : Any ) -> Union[str, Any]: pass def __snake_case ( self : Tuple , lowerCamelCase : Any , lowerCamelCase : str , lowerCamelCase : Union[str, Any]=False ) -> List[str]: __snake_case : Union[str, Any] = super()._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> Any: if not self.model_tester.is_training: return __snake_case , __snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : str = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(lowerCamelCase ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue __snake_case : List[str] = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.train() __snake_case : str = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) __snake_case : Optional[Any] = model(**lowerCamelCase ).loss loss.backward() def __snake_case ( self : List[Any] ) -> Any: __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return __snake_case : List[str] = False __snake_case : List[str] = True for model_class in self.all_model_classes: if model_class in get_values(lowerCamelCase ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue __snake_case : int = model_class(lowerCamelCase ) model.gradient_checkpointing_enable() model.to(lowerCamelCase ) model.train() __snake_case : List[Any] = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) __snake_case : Tuple = model(**lowerCamelCase ).loss loss.backward() def __snake_case ( self : List[Any] ) -> Dict: __snake_case , __snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Any = [ {"title": "multi_label_classification", "num_labels": 2, "dtype": torch.float}, {"title": "single_label_classification", "num_labels": 1, "dtype": torch.long}, {"title": "regression", "num_labels": 1, "dtype": torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(lowerCamelCase ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F'Testing {model_class} with {problem_type["title"]}' ): __snake_case : List[str] = problem_type["title"] __snake_case : Any = problem_type["num_labels"] __snake_case : Dict = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.train() __snake_case : str = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) if problem_type["num_labels"] > 1: __snake_case : str = inputs["labels"].unsqueeze(1 ).repeat(1 , problem_type["num_labels"] ) __snake_case : Dict = inputs["labels"].to(problem_type["dtype"] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=lowerCamelCase ) as warning_list: __snake_case : Optional[Any] = model(**lowerCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F'Something is going wrong in the regression problem: intercepted {w.message}' ) loss.backward() @slow def __snake_case ( self : Tuple ) -> Optional[int]: for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Any = LevitModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : Tuple ) -> str: return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __snake_case ( self : List[Any] ) -> Any: __snake_case : Optional[Any] = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( lowerCamelCase ) __snake_case : Dict = self.default_image_processor __snake_case : Tuple = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : int = model(**lowerCamelCase ) # verify the logits __snake_case : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([1.04_48, -0.37_45, -1.83_17] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) )
81
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
1
import os from argparse import ArgumentParser from typing import List import torch.utils.data from datasets import Dataset, IterableDataset from datasets.distributed import split_dataset_by_node _snake_case : Tuple = 4 _snake_case : Tuple = 3 class a (_lowerCAmelCase ): """simple docstring""" pass def lowerCAmelCase_ ( __lowerCamelCase ): for shard in shards: for i in range(__lowerCamelCase ): yield {"i": i, "shard": shard} def lowerCAmelCase_ ( ): __snake_case : Optional[Any] = int(os.environ["RANK"] ) __snake_case : Any = int(os.environ["WORLD_SIZE"] ) __snake_case : List[str] = ArgumentParser() parser.add_argument("--streaming" , type=__lowerCamelCase ) parser.add_argument("--local_rank" , type=__lowerCamelCase ) parser.add_argument("--num_workers" , type=__lowerCamelCase , default=0 ) __snake_case : Optional[Any] = parser.parse_args() __snake_case : List[Any] = args.streaming __snake_case : List[str] = args.num_workers __snake_case : int = {"shards": [F'shard_{shard_idx}' for shard_idx in range(__lowerCamelCase )]} __snake_case : List[Any] = IterableDataset.from_generator(__lowerCamelCase , gen_kwargs=__lowerCamelCase ) if not streaming: __snake_case : int = Dataset.from_list(list(__lowerCamelCase ) ) __snake_case : int = split_dataset_by_node(__lowerCamelCase , rank=__lowerCamelCase , world_size=__lowerCamelCase ) __snake_case : Optional[int] = torch.utils.data.DataLoader(__lowerCamelCase , num_workers=__lowerCamelCase ) __snake_case : Optional[Any] = NUM_SHARDS * NUM_ITEMS_PER_SHARD __snake_case : int = full_size // world_size expected_local_size += int(rank < (full_size % world_size) ) __snake_case : int = sum(1 for _ in dataloader ) if local_size != expected_local_size: raise FailedTestError(F'local_size {local_size} != expected_local_size {expected_local_size}' ) if __name__ == "__main__": main()
81
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
1
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
1
import numpy as np _snake_case : str = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class a : """simple docstring""" def __init__( self : Optional[int] ) -> None: __snake_case : Optional[int] = np.array(lowerCamelCase ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : str ) -> np.ndarray: __snake_case , __snake_case : Optional[int] = np.where(letter == self.SQUARE ) __snake_case : Union[str, Any] = np.concatenate([indexa + 1, indexa + 1] ) return indexes def __snake_case ( self : Union[str, Any] , lowerCamelCase : int , lowerCamelCase : int ) -> str: __snake_case : Optional[int] = self.SQUARE[indexa - 1, indexa - 1] return letter def __snake_case ( self : Union[str, Any] , lowerCamelCase : str ) -> str: __snake_case : Dict = message.lower() __snake_case : List[Any] = message.replace(" " , "" ) __snake_case : List[Any] = message.replace("j" , "i" ) __snake_case : Optional[Any] = np.empty((2, len(lowerCamelCase )) ) for letter_index in range(len(lowerCamelCase ) ): __snake_case : List[Any] = self.letter_to_numbers(message[letter_index] ) __snake_case : Any = numbers[0] __snake_case : Dict = numbers[1] __snake_case : Optional[Any] = first_step.reshape(2 * len(lowerCamelCase ) ) __snake_case : str = "" for numbers_index in range(len(lowerCamelCase ) ): __snake_case : str = int(second_step[numbers_index * 2] ) __snake_case : List[Any] = int(second_step[(numbers_index * 2) + 1] ) __snake_case : Optional[Any] = self.numbers_to_letter(lowerCamelCase , lowerCamelCase ) __snake_case : Union[str, Any] = encoded_message + letter return encoded_message def __snake_case ( self : Any , lowerCamelCase : str ) -> str: __snake_case : Tuple = message.lower() message.replace(" " , "" ) __snake_case : Dict = np.empty(2 * len(lowerCamelCase ) ) for letter_index in range(len(lowerCamelCase ) ): __snake_case : str = self.letter_to_numbers(message[letter_index] ) __snake_case : Dict = numbers[0] __snake_case : Union[str, Any] = numbers[1] __snake_case : int = first_step.reshape((2, len(lowerCamelCase )) ) __snake_case : List[Any] = "" for numbers_index in range(len(lowerCamelCase ) ): __snake_case : List[Any] = int(second_step[0, numbers_index] ) __snake_case : Optional[Any] = int(second_step[1, numbers_index] ) __snake_case : str = self.numbers_to_letter(lowerCamelCase , lowerCamelCase ) __snake_case : List[Any] = decoded_message + letter return decoded_message
81
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Any = str(__lowerCamelCase ) return n == n[::-1] def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : Optional[int] = 0 for i in range(1 , __lowerCamelCase ): if is_palindrome(__lowerCamelCase ) and is_palindrome(bin(__lowerCamelCase ).split("b" )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
81
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
1
def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = len(__lowerCamelCase ) __snake_case : Union[str, Any] = len(__lowerCamelCase ) __snake_case : Optional[Any] = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] __snake_case : int = True for i in range(__lowerCamelCase ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: __snake_case : Any = True if a[i].islower(): __snake_case : List[str] = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
81
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
def lowerCAmelCase_ ( __lowerCamelCase = 1_0 ): if not isinstance(__lowerCamelCase , __lowerCamelCase ) or n < 0: raise ValueError("Invalid input" ) __snake_case : List[str] = 1_0**n __snake_case : Optional[int] = 2_8_4_3_3 * (pow(2 , 7_8_3_0_4_5_7 , __lowerCamelCase )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(f'''{solution(10) = }''')
81
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
1
def lowerCAmelCase_ ( ): __snake_case : int = [] __snake_case : str = 1 while len(__lowerCamelCase ) < 1e6: constant.append(str(__lowerCamelCase ) ) i += 1 __snake_case : List[Any] = "".join(__lowerCamelCase ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[9_9] ) * int(constant[9_9_9] ) * int(constant[9_9_9_9] ) * int(constant[9_9_9_9_9] ) * int(constant[9_9_9_9_9_9] ) ) if __name__ == "__main__": print(solution())
81
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
1
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor _snake_case : Union[str, Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : str , **lowerCamelCase : Any ) -> None: warnings.warn( "The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use MobileViTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
1
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from argparse import ArgumentParser from accelerate.commands.config import get_config_parser from accelerate.commands.env import env_command_parser from accelerate.commands.launch import launch_command_parser from accelerate.commands.test import test_command_parser from accelerate.commands.tpu import tpu_command_parser def lowerCAmelCase_ ( ): __snake_case : int = ArgumentParser("Accelerate CLI tool" , usage="accelerate <command> [<args>]" , allow_abbrev=__lowerCamelCase ) __snake_case : Dict = parser.add_subparsers(help="accelerate command helpers" ) # Register commands get_config_parser(subparsers=__lowerCamelCase ) env_command_parser(subparsers=__lowerCamelCase ) launch_command_parser(subparsers=__lowerCamelCase ) tpu_command_parser(subparsers=__lowerCamelCase ) test_command_parser(subparsers=__lowerCamelCase ) # Let's go __snake_case : Any = parser.parse_args() if not hasattr(__lowerCamelCase , "func" ): parser.print_help() exit(1 ) # Run args.func(__lowerCamelCase ) if __name__ == "__main__": main()
81
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
1
import unittest import torch from diffusers import DDIMScheduler, DDPMScheduler, UNetaDModel from diffusers.training_utils import set_seed from diffusers.utils.testing_utils import slow _snake_case : Optional[Any] = False class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Optional[Any]=32 ) -> Dict: set_seed(0 ) __snake_case : int = UNetaDModel(sample_size=lowerCamelCase , in_channels=3 , out_channels=3 ) __snake_case : List[Any] = torch.optim.SGD(model.parameters() , lr=0.00_01 ) return model, optimizer @slow def __snake_case ( self : str ) -> List[Any]: __snake_case : int = "cpu" # ensure full determinism without setting the CUBLAS_WORKSPACE_CONFIG env variable __snake_case : Optional[Any] = DDPMScheduler( num_train_timesteps=1000 , beta_start=0.00_01 , beta_end=0.02 , beta_schedule="linear" , clip_sample=lowerCamelCase , ) __snake_case : Tuple = DDIMScheduler( num_train_timesteps=1000 , beta_start=0.00_01 , beta_end=0.02 , beta_schedule="linear" , clip_sample=lowerCamelCase , ) assert ddpm_scheduler.config.num_train_timesteps == ddim_scheduler.config.num_train_timesteps # shared batches for DDPM and DDIM set_seed(0 ) __snake_case : Tuple = [torch.randn((4, 3, 32, 32) ).clip(-1 , 1 ).to(lowerCamelCase ) for _ in range(4 )] __snake_case : str = [torch.randn((4, 3, 32, 32) ).to(lowerCamelCase ) for _ in range(4 )] __snake_case : List[Any] = [torch.randint(0 , 1000 , (4,) ).long().to(lowerCamelCase ) for _ in range(4 )] # train with a DDPM scheduler __snake_case , __snake_case : str = self.get_model_optimizer(resolution=32 ) model.train().to(lowerCamelCase ) for i in range(4 ): optimizer.zero_grad() __snake_case : Optional[Any] = ddpm_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) __snake_case : List[str] = model(lowerCamelCase , timesteps[i] ).sample __snake_case : Tuple = torch.nn.functional.mse_loss(lowerCamelCase , noise[i] ) loss.backward() optimizer.step() del model, optimizer # recreate the model and optimizer, and retry with DDIM __snake_case , __snake_case : Optional[int] = self.get_model_optimizer(resolution=32 ) model.train().to(lowerCamelCase ) for i in range(4 ): optimizer.zero_grad() __snake_case : str = ddim_scheduler.add_noise(clean_images[i] , noise[i] , timesteps[i] ) __snake_case : Any = model(lowerCamelCase , timesteps[i] ).sample __snake_case : Tuple = torch.nn.functional.mse_loss(lowerCamelCase , noise[i] ) loss.backward() optimizer.step() del model, optimizer self.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-5 ) ) self.assertTrue(torch.allclose(lowerCamelCase , lowerCamelCase , atol=1E-5 ) )
81
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
1
import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class a (pl.LightningModule ): """simple docstring""" def __init__( self : List[str] , lowerCamelCase : List[str] ) -> Any: super().__init__() __snake_case : Any = model __snake_case : List[str] = 2 __snake_case : List[Any] = nn.Linear(self.model.config.hidden_size , self.num_labels ) def __snake_case ( self : List[Any] ) -> List[Any]: pass def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): # load longformer model from model identifier __snake_case : List[str] = LongformerModel.from_pretrained(__lowerCamelCase ) __snake_case : Union[str, Any] = LightningModel(__lowerCamelCase ) __snake_case : Dict = torch.load(__lowerCamelCase , map_location=torch.device("cpu" ) ) lightning_model.load_state_dict(ckpt["state_dict"] ) # init longformer question answering model __snake_case : List[str] = LongformerForQuestionAnswering.from_pretrained(__lowerCamelCase ) # transfer weights longformer_for_qa.longformer.load_state_dict(lightning_model.model.state_dict() ) longformer_for_qa.qa_outputs.load_state_dict(lightning_model.qa_outputs.state_dict() ) longformer_for_qa.eval() # save model longformer_for_qa.save_pretrained(__lowerCamelCase ) print(F'Conversion successful. Model saved under {pytorch_dump_folder_path}' ) if __name__ == "__main__": _snake_case : int = argparse.ArgumentParser() # Required parameters parser.add_argument( "--longformer_model", default=None, type=str, required=True, help="model identifier of longformer. Should be either `longformer-base-4096` or `longformer-large-4096`.", ) parser.add_argument( "--longformer_question_answering_ckpt_path", default=None, type=str, required=True, help="Path the official PyTorch Lightning Checkpoint.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) _snake_case : str = parser.parse_args() convert_longformer_qa_checkpoint_to_pytorch( args.longformer_model, args.longformer_question_answering_ckpt_path, args.pytorch_dump_folder_path )
81
import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case , __snake_case : List[Any] = image.size __snake_case , __snake_case : Tuple = (x - x % 3_2 for x in (w, h)) # resize to integer multiple of 32 __snake_case : str = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case : int = np.array(__lowerCamelCase ).astype(np.floataa ) / 2_5_5.0 __snake_case : Union[str, Any] = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case : Union[str, Any] = torch.from_numpy(__lowerCamelCase ) return 2.0 * image - 1.0 class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : VQModel , lowerCamelCase : UNetaDModel , lowerCamelCase : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ) -> Union[str, Any]: super().__init__() self.register_modules(vqvae=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase ) @torch.no_grad() def __call__( self : List[str] , lowerCamelCase : Union[torch.Tensor, PIL.Image.Image] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : Optional[int] = 100 , lowerCamelCase : Optional[float] = 0.0 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> Union[Tuple, ImagePipelineOutput]: if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : Any = 1 elif isinstance(lowerCamelCase , torch.Tensor ): __snake_case : Any = image.shape[0] else: raise ValueError(F'`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(lowerCamelCase )}' ) if isinstance(lowerCamelCase , PIL.Image.Image ): __snake_case : List[Any] = preprocess(lowerCamelCase ) __snake_case , __snake_case : int = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case : str = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case : str = next(self.unet.parameters() ).dtype __snake_case : Tuple = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=self.device , dtype=lowerCamelCase ) __snake_case : List[Any] = image.to(device=self.device , dtype=lowerCamelCase ) # set timesteps and move to the correct device self.scheduler.set_timesteps(lowerCamelCase , device=self.device ) __snake_case : str = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case : Union[str, Any] = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case : int = {} if accepts_eta: __snake_case : List[str] = eta for t in self.progress_bar(lowerCamelCase ): # concat latents and low resolution image in the channel dimension. __snake_case : Union[str, Any] = torch.cat([latents, image] , dim=1 ) __snake_case : Optional[Any] = self.scheduler.scale_model_input(lowerCamelCase , lowerCamelCase ) # predict the noise residual __snake_case : int = self.unet(lowerCamelCase , lowerCamelCase ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case : Union[str, Any] = self.scheduler.step(lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ).prev_sample # decode the image latents with the VQVAE __snake_case : List[Any] = self.vqvae.decode(lowerCamelCase ).sample __snake_case : Dict = torch.clamp(lowerCamelCase , -1.0 , 1.0 ) __snake_case : Any = image / 2 + 0.5 __snake_case : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case : Tuple = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
1
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _snake_case : List[str] = logging.get_logger(__name__) # pylint: disable=invalid-name _snake_case : Optional[int] = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=8 ): __snake_case : List[Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 __snake_case : Optional[int] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : DDPMScheduler , lowerCamelCase : VQModel , ) -> Optional[Any]: super().__init__() self.register_modules( unet=lowerCamelCase , scheduler=lowerCamelCase , movq=lowerCamelCase , ) __snake_case : int = 2 ** (len(self.movq.config.block_out_channels ) - 1) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : int , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[int] ) -> Optional[Any]: if latents is None: __snake_case : List[str] = randn_tensor(lowerCamelCase , generator=lowerCamelCase , device=lowerCamelCase , dtype=lowerCamelCase ) else: if latents.shape != shape: raise ValueError(F'Unexpected latents shape, got {latents.shape}, expected {shape}' ) __snake_case : Optional[int] = latents.to(lowerCamelCase ) __snake_case : List[Any] = latents * scheduler.init_noise_sigma return latents def __snake_case ( self : Optional[Any] , lowerCamelCase : Any=0 ) -> List[str]: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : List[str] = torch.device(F'cuda:{gpu_id}' ) __snake_case : List[Any] = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Optional[int] , lowerCamelCase : List[str]=0 ) -> str: if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) __snake_case : List[Any] = torch.device(F'cuda:{gpu_id}' ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowerCamelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) __snake_case : List[str] = None for cpu_offloaded_model in [self.unet, self.movq]: __snake_case , __snake_case : List[str] = cpu_offload_with_hook(lowerCamelCase , lowerCamelCase , prev_module_hook=lowerCamelCase ) # We'll offload the last model manually. __snake_case : Optional[int] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : Optional[int] ) -> Dict: if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowerCamelCase ) def __call__( self : Optional[Any] , lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , lowerCamelCase : Union[torch.FloatTensor, List[torch.FloatTensor]] , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 100 , lowerCamelCase : float = 4.0 , lowerCamelCase : int = 1 , lowerCamelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , ) -> str: __snake_case : Any = self._execution_device __snake_case : Optional[Any] = guidance_scale > 1.0 if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : str = torch.cat(lowerCamelCase , dim=0 ) __snake_case : Dict = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Any = torch.cat(lowerCamelCase , dim=0 ) if do_classifier_free_guidance: __snake_case : List[str] = image_embeds.repeat_interleave(lowerCamelCase , dim=0 ) __snake_case : Dict = negative_image_embeds.repeat_interleave(lowerCamelCase , dim=0 ) __snake_case : Any = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowerCamelCase ) self.scheduler.set_timesteps(lowerCamelCase , device=lowerCamelCase ) __snake_case : Any = self.scheduler.timesteps __snake_case : Any = self.unet.config.in_channels __snake_case , __snake_case : Dict = downscale_height_and_width(lowerCamelCase , lowerCamelCase , self.movq_scale_factor ) # create initial latent __snake_case : Optional[Any] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowerCamelCase , lowerCamelCase , lowerCamelCase , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowerCamelCase ) ): # expand the latents if we are doing classifier free guidance __snake_case : str = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents __snake_case : Optional[int] = {"image_embeds": image_embeds} __snake_case : Any = self.unet( sample=lowerCamelCase , timestep=lowerCamelCase , encoder_hidden_states=lowerCamelCase , added_cond_kwargs=lowerCamelCase , return_dict=lowerCamelCase , )[0] if do_classifier_free_guidance: __snake_case , __snake_case : Union[str, Any] = noise_pred.split(latents.shape[1] , dim=1 ) __snake_case , __snake_case : Any = noise_pred.chunk(2 ) __snake_case , __snake_case : int = variance_pred.chunk(2 ) __snake_case : Tuple = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) __snake_case : List[Any] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): __snake_case , __snake_case : List[str] = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 __snake_case : List[str] = self.scheduler.step( lowerCamelCase , lowerCamelCase , lowerCamelCase , generator=lowerCamelCase , )[0] # post-processing __snake_case : List[str] = self.movq.decode(lowerCamelCase , force_not_quantize=lowerCamelCase )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}' ) if output_type in ["np", "pil"]: __snake_case : List[str] = image * 0.5 + 0.5 __snake_case : List[str] = image.clamp(0 , 1 ) __snake_case : List[str] = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": __snake_case : int = self.numpy_to_pil(lowerCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCamelCase )
81
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _snake_case : Optional[int] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerOnnxConfig", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : str = ["PoolFormerFeatureExtractor"] _snake_case : List[Any] = ["PoolFormerImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : str = [ "POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "PoolFormerForImageClassification", "PoolFormerModel", "PoolFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure)
81
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = [2, 2, 6, 2] if "tiny" in model_name else [2, 2, 1_8, 2] __snake_case : Optional[int] = True if "large" in model_name or "huge" in model_name else False __snake_case : List[Any] = True if "large" in model_name or "huge" in model_name else False __snake_case : List[Any] = True if "large" in model_name or "huge" in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: __snake_case : Optional[Any] = [3, 3, 3, 3] __snake_case : Optional[Any] = [5, 5, 5, 5] elif "fl4" in model_name: __snake_case : List[Any] = [4, 4, 4, 4] __snake_case : Any = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: __snake_case : List[str] = [3, 3, 3, 3] if "lrf" in model_name: __snake_case : Any = [3, 3, 3, 3] else: __snake_case : str = [2, 2, 2, 2] if "tiny" in model_name: __snake_case : int = 9_6 elif "small" in model_name: __snake_case : Dict = 9_6 elif "base" in model_name: __snake_case : Any = 1_2_8 elif "large" in model_name: __snake_case : List[str] = 1_9_2 elif "xlarge" in model_name: __snake_case : List[str] = 2_5_6 elif "huge" in model_name: __snake_case : List[str] = 3_5_2 # set label information __snake_case : List[Any] = "huggingface/label-files" if "large" in model_name or "huge" in model_name: __snake_case : Any = "imagenet-22k-id2label.json" else: __snake_case : str = "imagenet-1k-id2label.json" __snake_case : Optional[Any] = json.load(open(hf_hub_download(__lowerCamelCase , __lowerCamelCase , repo_type="dataset" ) , "r" ) ) __snake_case : Any = {int(__lowerCamelCase ): v for k, v in idalabel.items()} __snake_case : Optional[Any] = {v: k for k, v in idalabel.items()} __snake_case : Dict = FocalNetConfig( embed_dim=__lowerCamelCase , depths=__lowerCamelCase , focal_levels=__lowerCamelCase , focal_windows=__lowerCamelCase , use_conv_embed=__lowerCamelCase , idalabel=__lowerCamelCase , labelaid=__lowerCamelCase , use_post_layernorm=__lowerCamelCase , use_layerscale=__lowerCamelCase , ) return config def lowerCAmelCase_ ( __lowerCamelCase ): if "patch_embed.proj" in name: __snake_case : Union[str, Any] = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: __snake_case : Optional[Any] = name.replace("patch_embed.norm" , "embeddings.norm" ) if "layers" in name: __snake_case : Optional[int] = "encoder." + name if "encoder.layers" in name: __snake_case : List[str] = name.replace("encoder.layers" , "encoder.stages" ) if "downsample.proj" in name: __snake_case : Optional[Any] = name.replace("downsample.proj" , "downsample.projection" ) if "blocks" in name: __snake_case : Dict = name.replace("blocks" , "layers" ) if "modulation.f.weight" in name or "modulation.f.bias" in name: __snake_case : Tuple = name.replace("modulation.f" , "modulation.projection_in" ) if "modulation.h.weight" in name or "modulation.h.bias" in name: __snake_case : Union[str, Any] = name.replace("modulation.h" , "modulation.projection_context" ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: __snake_case : Any = name.replace("modulation.proj" , "modulation.projection_out" ) if name == "norm.weight": __snake_case : Any = "layernorm.weight" if name == "norm.bias": __snake_case : Optional[int] = "layernorm.bias" if "head" in name: __snake_case : Union[str, Any] = name.replace("head" , "classifier" ) else: __snake_case : Union[str, Any] = "focalnet." + name return name def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=False ): # fmt: off __snake_case : str = { "focalnet-tiny": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth", "focalnet-tiny-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth", "focalnet-small": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth", "focalnet-small-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth", "focalnet-base": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth", "focalnet-base-lrf": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth", "focalnet-large-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth", "focalnet-large-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth", "focalnet-xlarge-lrf-fl3": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth", "focalnet-xlarge-lrf-fl4": "https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth", } # fmt: on __snake_case : Union[str, Any] = model_name_to_url[model_name] print("Checkpoint URL: " , __lowerCamelCase ) __snake_case : Tuple = torch.hub.load_state_dict_from_url(__lowerCamelCase , map_location="cpu" )["model"] # rename keys for key in state_dict.copy().keys(): __snake_case : Tuple = state_dict.pop(__lowerCamelCase ) __snake_case : List[str] = val __snake_case : Any = get_focalnet_config(__lowerCamelCase ) __snake_case : Optional[Any] = FocalNetForImageClassification(__lowerCamelCase ) model.eval() # load state dict model.load_state_dict(__lowerCamelCase ) # verify conversion __snake_case : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" __snake_case : List[str] = BitImageProcessor( do_resize=__lowerCamelCase , size={"shortest_edge": 2_5_6} , resample=PILImageResampling.BILINEAR , do_center_crop=__lowerCamelCase , crop_size=2_2_4 , do_normalize=__lowerCamelCase , image_mean=__lowerCamelCase , image_std=__lowerCamelCase , ) __snake_case : Optional[int] = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) __snake_case : Tuple = processor(images=__lowerCamelCase , return_tensors="pt" ) __snake_case : Tuple = transforms.Compose( [ transforms.Resize(2_5_6 ), transforms.CenterCrop(2_2_4 ), transforms.ToTensor(), transforms.Normalize(mean=[0.4_8_5, 0.4_5_6, 0.4_0_6] , std=[0.2_2_9, 0.2_2_4, 0.2_2_5] ), ] ) __snake_case : Tuple = image_transforms(__lowerCamelCase ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values , __lowerCamelCase , atol=1e-4 ) __snake_case : Union[str, Any] = model(**__lowerCamelCase ) __snake_case : List[str] = outputs.logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) print("First values of logits:" , outputs.logits[0, :3] ) if model_name == "focalnet-tiny": __snake_case : List[Any] = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ) elif model_name == "focalnet-tiny-lrf": __snake_case : Union[str, Any] = torch.tensor([1.1_6_6_9, 0.0_1_2_5, -0.1_6_9_5] ) elif model_name == "focalnet-small": __snake_case : List[Any] = torch.tensor([0.4_9_1_7, -0.0_4_3_0, 0.1_3_4_1] ) elif model_name == "focalnet-small-lrf": __snake_case : List[str] = torch.tensor([-0.2_5_8_8, -0.5_3_4_2, -0.2_3_3_1] ) elif model_name == "focalnet-base": __snake_case : List[str] = torch.tensor([-0.1_6_5_5, -0.4_0_9_0, -0.1_7_3_0] ) elif model_name == "focalnet-base-lrf": __snake_case : Tuple = torch.tensor([0.5_3_0_6, -0.0_4_8_3, -0.3_9_2_8] ) assert torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(F'Saving model and processor of {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(__lowerCamelCase ) processor.save_pretrained(__lowerCamelCase ) if push_to_hub: print(F'Pushing model and processor of {model_name} to the hub...' ) model.push_to_hub(F'{model_name}' ) processor.push_to_hub(F'{model_name}' ) if __name__ == "__main__": _snake_case : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="focalnet-tiny", type=str, help="Name of the FocalNet model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether to push the model and processor to the hub.", ) _snake_case : Dict = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
81
import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, SamImageProcessor, SamProcessor if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : Tuple ) -> Optional[Any]: __snake_case : Dict = tempfile.mkdtemp() __snake_case : Any = SamImageProcessor() __snake_case : Optional[int] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : Optional[Any] , **lowerCamelCase : Optional[int] ) -> Optional[Any]: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[Any] ) -> Dict: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : int ) -> List[Any]: __snake_case : List[str] = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : int = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : List[Any] ) -> Dict: __snake_case : int = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Dict = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Optional[Any] = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : List[str] ) -> Tuple: __snake_case : int = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Optional[int] = self.prepare_image_inputs() __snake_case : List[str] = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : Dict = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop original_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_torch def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Tuple = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[str] = [torch.ones((1, 3, 5, 5) )] __snake_case : Tuple = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : int = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , torch.tensor(lowerCamelCase ) , torch.tensor(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : List[str] = [np.ones((1, 3, 5, 5) )] __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : str = [[1, 0], [0, 1]] with self.assertRaises(lowerCamelCase ): __snake_case : Optional[int] = processor.post_process_masks(lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) ) @require_vision @require_tf class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : int = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : str , **lowerCamelCase : Any ) -> Tuple: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> Any: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : str ) -> List[Any]: __snake_case : int = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : Dict = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def __snake_case ( self : int ) -> List[str]: __snake_case : List[Any] = SamProcessor(image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) __snake_case : Optional[Any] = self.get_image_processor(do_normalize=lowerCamelCase , padding_value=1.0 ) __snake_case : Tuple = SamProcessor.from_pretrained(self.tmpdirname , do_normalize=lowerCamelCase , padding_value=1.0 ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: __snake_case : str = self.get_image_processor() __snake_case : Union[str, Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : int = image_processor(lowerCamelCase , return_tensors="np" ) __snake_case : List[str] = processor(images=lowerCamelCase , return_tensors="np" ) input_feat_extract.pop("original_sizes" ) # pop original_sizes as it is popped in the processor input_feat_extract.pop("reshaped_input_sizes" ) # pop reshaped_input_sizes as it is popped in the processor for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) @require_tf def __snake_case ( self : Any ) -> Optional[int]: __snake_case : List[str] = self.get_image_processor() __snake_case : Dict = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Union[str, Any] = [tf.ones((1, 3, 5, 5) )] __snake_case : List[Any] = [[1764, 2646]] __snake_case : Dict = [[683, 1024]] __snake_case : List[str] = processor.post_process_masks(lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Optional[Any] = processor.post_process_masks( lowerCamelCase , tf.convert_to_tensor(lowerCamelCase ) , tf.convert_to_tensor(lowerCamelCase ) , return_tensors="tf" , ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) # should also work with np __snake_case : Union[str, Any] = [np.ones((1, 3, 5, 5) )] __snake_case : List[str] = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) self.assertEqual(masks[0].shape , (1, 3, 1764, 2646) ) __snake_case : Tuple = [[1, 0], [0, 1]] with self.assertRaises(tf.errors.InvalidArgumentError ): __snake_case : Dict = processor.post_process_masks( lowerCamelCase , np.array(lowerCamelCase ) , np.array(lowerCamelCase ) , return_tensors="tf" ) @require_vision @require_torchvision class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : List[str] ) -> str: __snake_case : Optional[int] = tempfile.mkdtemp() __snake_case : str = SamImageProcessor() __snake_case : List[Any] = SamProcessor(lowerCamelCase ) processor.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str] , **lowerCamelCase : Any ) -> Dict: return AutoProcessor.from_pretrained(self.tmpdirname , **lowerCamelCase ).image_processor def __snake_case ( self : Optional[int] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : str = [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] __snake_case : List[Any] = [Image.fromarray(np.moveaxis(lowerCamelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs @is_pt_tf_cross_test def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : str = self.get_image_processor() __snake_case : str = SamProcessor(image_processor=lowerCamelCase ) __snake_case : List[Any] = np.random.randint(0 , 2 , size=(1, 3, 5, 5) ).astype(np.floataa ) __snake_case : Dict = [tf.convert_to_tensor(lowerCamelCase )] __snake_case : List[Any] = [torch.tensor(lowerCamelCase )] __snake_case : Optional[Any] = [[1764, 2646]] __snake_case : Optional[int] = [[683, 1024]] __snake_case : Union[str, Any] = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="tf" ) __snake_case : Dict = processor.post_process_masks( lowerCamelCase , lowerCamelCase , lowerCamelCase , return_tensors="pt" ) self.assertTrue(np.all(tf_masks[0].numpy() == pt_masks[0].numpy() ) ) @is_pt_tf_cross_test def __snake_case ( self : List[Any] ) -> List[str]: __snake_case : Any = self.get_image_processor() __snake_case : List[Any] = SamProcessor(image_processor=lowerCamelCase ) __snake_case : Dict = self.prepare_image_inputs() __snake_case : Any = image_processor(lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Optional[Any] = processor(images=lowerCamelCase , return_tensors="pt" )["pixel_values"].numpy() __snake_case : Tuple = image_processor(lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() __snake_case : List[Any] = processor(images=lowerCamelCase , return_tensors="tf" )["pixel_values"].numpy() self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) ) self.assertTrue(np.allclose(lowerCamelCase , lowerCamelCase ) )
81
1
from __future__ import annotations _snake_case : Any = "Muhammad Umer Farooq" _snake_case : List[str] = "MIT" _snake_case : Tuple = "1.0.0" _snake_case : Any = "Muhammad Umer Farooq" _snake_case : Optional[Any] = "contact@muhammadumerfarooq.me" _snake_case : Optional[Any] = "Alpha" import re from html.parser import HTMLParser from urllib import parse import requests class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : str , lowerCamelCase : str ) -> None: super().__init__() __snake_case : list[str] = [] __snake_case : Optional[Any] = domain def __snake_case ( self : str , lowerCamelCase : str , lowerCamelCase : list[tuple[str, str | None]] ) -> None: # Only parse the 'anchor' tag. if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: __snake_case : Any = parse.urljoin(self.domain , lowerCamelCase ) self.urls.append(lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): return ".".join(get_sub_domain_name(__lowerCamelCase ).split("." )[-2:] ) def lowerCAmelCase_ ( __lowerCamelCase ): return parse.urlparse(__lowerCamelCase ).netloc def lowerCAmelCase_ ( __lowerCamelCase = "https://github.com" ): __snake_case : int = get_domain_name(__lowerCamelCase ) # Initialize the parser __snake_case : List[str] = Parser(__lowerCamelCase ) try: # Open URL __snake_case : Dict = requests.get(__lowerCamelCase ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through __snake_case : int = set() for link in parser.urls: # open URL. # read = requests.get(link) try: __snake_case : Dict = requests.get(__lowerCamelCase ) # Get the valid email. __snake_case : int = re.findall("[a-zA-Z0-9]+@" + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(__lowerCamelCase ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(__lowerCamelCase ) if __name__ == "__main__": _snake_case : str = emails_from_url("https://github.com") print(f'''{len(emails)} emails found:''') print("\n".join(sorted(emails)))
81
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Any , *lowerCamelCase : int , **lowerCamelCase : int ) -> None: warnings.warn( "The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use PerceiverImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, is_vision_available, ) _snake_case : Union[str, Any] = { "configuration_owlvit": [ "OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "OwlViTConfig", "OwlViTOnnxConfig", "OwlViTTextConfig", "OwlViTVisionConfig", ], "processing_owlvit": ["OwlViTProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[Any] = ["OwlViTFeatureExtractor"] _snake_case : Optional[int] = ["OwlViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST", "OwlViTModel", "OwlViTPreTrainedModel", "OwlViTTextModel", "OwlViTVisionModel", "OwlViTForObjectDetection", ] if TYPE_CHECKING: from .configuration_owlvit import ( OWLVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, OwlViTConfig, OwlViTOnnxConfig, OwlViTTextConfig, OwlViTVisionConfig, ) from .processing_owlvit import OwlViTProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_owlvit import OwlViTFeatureExtractor from .image_processing_owlvit import OwlViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_owlvit import ( OWLVIT_PRETRAINED_MODEL_ARCHIVE_LIST, OwlViTForObjectDetection, OwlViTModel, OwlViTPreTrainedModel, OwlViTTextModel, OwlViTVisionModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
from numpy import exp, pi, sqrt def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase = 0.0 , __lowerCamelCase = 1.0 ): return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
81
from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : int = logging.get_logger(__name__) _snake_case : str = { "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[Any] = "van" def __init__( self : Optional[int] , lowerCamelCase : Any=224 , lowerCamelCase : str=3 , lowerCamelCase : Any=[7, 3, 3, 3] , lowerCamelCase : Dict=[4, 2, 2, 2] , lowerCamelCase : List[Any]=[64, 128, 320, 512] , lowerCamelCase : str=[3, 3, 12, 3] , lowerCamelCase : Dict=[8, 8, 4, 4] , lowerCamelCase : Any="gelu" , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Tuple=1E-6 , lowerCamelCase : Optional[int]=1E-2 , lowerCamelCase : int=0.0 , lowerCamelCase : Optional[Any]=0.0 , **lowerCamelCase : Optional[int] , ) -> int: super().__init__(**lowerCamelCase ) __snake_case : Union[str, Any] = image_size __snake_case : Any = num_channels __snake_case : Any = patch_sizes __snake_case : List[Any] = strides __snake_case : str = hidden_sizes __snake_case : Dict = depths __snake_case : Optional[int] = mlp_ratios __snake_case : Dict = hidden_act __snake_case : Union[str, Any] = initializer_range __snake_case : List[str] = layer_norm_eps __snake_case : Optional[int] = layer_scale_init_value __snake_case : List[Any] = drop_path_rate __snake_case : int = dropout_rate
81
1
import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class a : """simple docstring""" def __snake_case ( self : Tuple , lowerCamelCase : Dict , lowerCamelCase : Tuple , lowerCamelCase : Optional[Any] ) -> Dict: return None class a : """simple docstring""" def __snake_case ( self : Optional[Any] , lowerCamelCase : Dict , lowerCamelCase : Any , lowerCamelCase : str , lowerCamelCase : Tuple ) -> List[Any]: return None class a (unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = [ # (model_name, model_kwargs) ("bert-base-cased", {}), ("gpt2", {"use_cache": False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def __snake_case ( self : Dict ) -> Any: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase , "tf" , 12 , **lowerCamelCase ) @require_torch @slow def __snake_case ( self : List[Any] ) -> Dict: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase , "pt" , 12 , **lowerCamelCase ) @require_torch @slow def __snake_case ( self : int ) -> Tuple: from transformers import BertModel __snake_case : int = ["[UNK]", "[SEP]", "[CLS]", "[PAD]", "[MASK]", "some", "other", "words"] with NamedTemporaryFile(mode="w+t" ) as vocab_file: vocab_file.write("\n".join(lowerCamelCase ) ) vocab_file.flush() __snake_case : List[Any] = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: __snake_case : List[str] = BertModel(BertConfig(vocab_size=len(lowerCamelCase ) ) ) model.save_pretrained(lowerCamelCase ) self._test_export(lowerCamelCase , "pt" , 12 , lowerCamelCase ) @require_tf @slow def __snake_case ( self : Dict ) -> Tuple: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: __snake_case : int = self._test_export(lowerCamelCase , "tf" , 12 , **lowerCamelCase ) __snake_case : List[Any] = quantize(Path(lowerCamelCase ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase ).stat().st_size: self.fail("Quantized model is bigger than initial ONNX model" ) @require_torch @slow def __snake_case ( self : str ) -> Optional[Any]: for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: __snake_case : Union[str, Any] = self._test_export(lowerCamelCase , "pt" , 12 , **lowerCamelCase ) __snake_case : Optional[Any] = quantize(lowerCamelCase ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase ).stat().st_size: self.fail("Quantized model is bigger than initial ONNX model" ) def __snake_case ( self : Tuple , lowerCamelCase : int , lowerCamelCase : Dict , lowerCamelCase : Tuple , lowerCamelCase : Tuple=None , **lowerCamelCase : List[str] ) -> Dict: try: # Compute path with TemporaryDirectory() as tempdir: __snake_case : Tuple = Path(lowerCamelCase ).joinpath("model.onnx" ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase ) return path except Exception as e: self.fail(lowerCamelCase ) @require_torch @require_tokenizers @slow def __snake_case ( self : List[Any] ) -> List[Any]: from transformers import BertModel __snake_case : Optional[int] = BertModel(BertConfig.from_pretrained("lysandre/tiny-bert-random" ) ) __snake_case : Dict = BertTokenizerFast.from_pretrained("lysandre/tiny-bert-random" ) self._test_infer_dynamic_axis(lowerCamelCase , lowerCamelCase , "pt" ) @require_tf @require_tokenizers @slow def __snake_case ( self : Any ) -> Optional[int]: from transformers import TFBertModel __snake_case : str = TFBertModel(BertConfig.from_pretrained("lysandre/tiny-bert-random" ) ) __snake_case : List[Any] = BertTokenizerFast.from_pretrained("lysandre/tiny-bert-random" ) self._test_infer_dynamic_axis(lowerCamelCase , lowerCamelCase , "tf" ) def __snake_case ( self : Optional[int] , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : str ) -> Any: __snake_case : Any = FeatureExtractionPipeline(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[int] = ["input_ids", "token_type_ids", "attention_mask", "output_0", "output_1"] __snake_case , __snake_case , __snake_case , __snake_case : Optional[int] = infer_shapes(lowerCamelCase , lowerCamelCase ) # Assert all variables are present self.assertEqual(len(lowerCamelCase ) , len(lowerCamelCase ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , lowerCamelCase ) self.assertSequenceEqual(variable_names[3:] , lowerCamelCase ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: "batch", 1: "sequence"} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes["output_0"] , {0: "batch", 1: "sequence"} ) self.assertDictEqual(shapes["output_1"] , {0: "batch"} ) def __snake_case ( self : Dict ) -> Tuple: __snake_case : List[str] = ["input_ids", "attention_mask", "token_type_ids"] __snake_case : Dict = {"input_ids": [1, 2, 3, 4], "attention_mask": [0, 0, 0, 0], "token_type_ids": [1, 1, 1, 1]} __snake_case , __snake_case : str = ensure_valid_input(FuncContiguousArgs() , lowerCamelCase , lowerCamelCase ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(lowerCamelCase ) , 3 ) # Should have exactly the same input names self.assertEqual(set(lowerCamelCase ) , set(lowerCamelCase ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(lowerCamelCase , (tokens["input_ids"], tokens["token_type_ids"], tokens["attention_mask"]) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) __snake_case , __snake_case : Tuple = ensure_valid_input(FuncNonContiguousArgs() , lowerCamelCase , lowerCamelCase ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(lowerCamelCase ) , 1 ) self.assertEqual(len(lowerCamelCase ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens["input_ids"] ) self.assertEqual(ordered_input_names[0] , "input_ids" ) def __snake_case ( self : int ) -> Dict: __snake_case : Any = generate_identified_filename(Path("/home/something/my_fake_model.onnx" ) , "-test" ) self.assertEqual("/home/something/my_fake_model-test.onnx" , generated.as_posix() )
81
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _snake_case : Union[str, Any] = logging.getLogger(__name__) def lowerCAmelCase_ ( ): __snake_case : int = argparse.ArgumentParser( description="Prepare TFRecord shards from pre-tokenized samples of the wikitext dataset." ) parser.add_argument( "--dataset_name" , type=__lowerCamelCase , default="wikitext" , help="Name of the training. Explore datasets at: hf.co/datasets." , ) parser.add_argument( "--dataset_config" , type=__lowerCamelCase , default="wikitext-103-raw-v1" , help="Configuration name of the dataset." ) parser.add_argument( "--tokenizer_name_or_path" , type=__lowerCamelCase , default="sayakpaul/unigram-tokenizer-wikitext" , help="Tokenizer identifier. Can be a local filepath or a Hub identifier." , ) parser.add_argument( "--shard_size" , type=__lowerCamelCase , default=1_0_0_0 , help="Number of entries to go in a single shard." , ) parser.add_argument("--split" , type=__lowerCamelCase , default="train" , choices=["train", "test", "validation"] ) parser.add_argument( "--limit" , default=__lowerCamelCase , type=__lowerCamelCase , help="Limit the number of shards (used for debugging)." , ) parser.add_argument( "--max_length" , type=__lowerCamelCase , default=5_1_2 , help="Maximum sequence length. For training on TPUs, it helps to have a maximum" " sequence length that is a multiple of 8." , ) parser.add_argument( "--output_dir" , default="tf-tpu" , type=__lowerCamelCase , help="Output directory where the TFRecord shards will be saved. If the" " path is appended with `gs://` ('gs://tf-tpu', for example) then the TFRecord" " shards will be directly saved to a Google Cloud Storage bucket." , ) __snake_case : List[str] = parser.parse_args() return args def lowerCAmelCase_ ( __lowerCamelCase ): def fn(__lowerCamelCase ): return tokenizer(examples["text"] ) return fn def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = [] for i in range(len(tokenized_data["input_ids"] ) ): __snake_case : Tuple = { "input_ids": tf.train.Feature(intaa_list=tf.train.IntaaList(value=tokenized_data["input_ids"][i] ) ), "attention_mask": tf.train.Feature( intaa_list=tf.train.IntaaList(value=tokenized_data["attention_mask"][i] ) ), } __snake_case : List[Any] = tf.train.Features(feature=__lowerCamelCase ) __snake_case : str = tf.train.Example(features=__lowerCamelCase ) __snake_case : List[str] = example.SerializeToString() records.append(__lowerCamelCase ) return records def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[int] = datasets.load_dataset(args.dataset_name , args.dataset_config , split=args.split ) if args.limit is not None: __snake_case : Optional[Any] = min(len(__lowerCamelCase ) , args.limit ) __snake_case : Dict = dataset.select(range(__lowerCamelCase ) ) print(F'Limiting the dataset to {args.limit} entries.' ) __snake_case : Dict = AutoTokenizer.from_pretrained(args.tokenizer_name_or_path ) # Handle output directory creation. # For serializing into a Google Cloud Storage Bucket, one needs to first # create a bucket. if "gs" not in args.output_dir: if not os.path.exists(args.output_dir ): os.makedirs(args.output_dir ) __snake_case : Dict = os.path.join(args.output_dir , args.split ) if not os.path.exists(__lowerCamelCase ): os.makedirs(__lowerCamelCase ) else: __snake_case : str = os.path.join(args.output_dir , args.split ) # Tokenize the whole dataset at once. __snake_case : Any = tokenize_function(__lowerCamelCase ) __snake_case : Optional[Any] = dataset.map(__lowerCamelCase , batched=__lowerCamelCase , num_proc=4 , remove_columns=["text"] ) # We need to concatenate all our texts together, and then split the result # into chunks of a fixed size, which we will call block_size. To do this, we # will use the map method again, with the option batched=True. When we use batched=True, # the function we pass to map() will be passed multiple inputs at once, allowing us # to group them into more or fewer examples than we had in the input. # This allows us to create our new fixed-length samples. The advantage of this # method is that we don't lose a whole lot of content from the dataset compared to the # case where we simply tokenize with a pre-defined max_length. def group_texts(__lowerCamelCase ): # Concatenate all texts. __snake_case : List[str] = {k: sum(examples[k] , [] ) for k in examples.keys()} __snake_case : List[Any] = len(concatenated_examples[list(examples.keys() )[0]] ) # We drop the small remainder, though you could add padding instead if the model supports it # In this, as in all things, we advise you to follow your heart 🫀 __snake_case : Any = (total_length // args.max_length) * args.max_length # Split by chunks of max_len. __snake_case : int = { k: [t[i : i + args.max_length] for i in range(0 , __lowerCamelCase , args.max_length )] for k, t in concatenated_examples.items() } return result __snake_case : Any = dataset_tokenized.map(__lowerCamelCase , batched=__lowerCamelCase , batch_size=1_0_0_0 , num_proc=4 ) __snake_case : Optional[Any] = 0 __snake_case : Optional[Any] = 0 for shard in range(0 , len(__lowerCamelCase ) , args.shard_size ): __snake_case : List[str] = grouped_dataset[shard : shard + args.shard_size] __snake_case : Any = len(dataset_snapshot["input_ids"] ) __snake_case : List[Any] = os.path.join(__lowerCamelCase , F'dataset-{shard_count}-{records_containing}.tfrecord' ) __snake_case : Optional[Any] = get_serialized_examples(__lowerCamelCase ) with tf.io.TFRecordWriter(__lowerCamelCase ) as out_file: for i in range(len(__lowerCamelCase ) ): __snake_case : Union[str, Any] = serialized_examples[i] out_file.write(__lowerCamelCase ) print("Wrote file {} containing {} records".format(__lowerCamelCase , __lowerCamelCase ) ) shard_count += 1 total_records += records_containing with open(F'split-{args.split}-records-count.txt' , "w" ) as f: print(F'Total {args.split} records: {total_records}' , file=__lowerCamelCase ) if __name__ == "__main__": _snake_case : List[Any] = parse_args() main(args)
81
1
import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class a : """simple docstring""" def __init__( self : Union[str, Any] , lowerCamelCase : Any , lowerCamelCase : Tuple=13 , lowerCamelCase : Dict=3 , lowerCamelCase : Tuple=True , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Any=0.1 , lowerCamelCase : str=0.1 , lowerCamelCase : Union[str, Any]=224 , lowerCamelCase : str=1000 , lowerCamelCase : int=[3, 3, 6, 4] , lowerCamelCase : List[str]=[48, 56, 112, 220] , ) -> List[Any]: __snake_case : Optional[int] = parent __snake_case : Tuple = batch_size __snake_case : Union[str, Any] = num_channels __snake_case : Optional[Any] = is_training __snake_case : List[str] = use_labels __snake_case : int = hidden_dropout_prob __snake_case : int = attention_probs_dropout_prob __snake_case : List[str] = num_labels __snake_case : Optional[Any] = image_size __snake_case : Dict = layer_depths __snake_case : List[Any] = embed_dims def __snake_case ( self : Optional[int] ) -> Tuple: __snake_case : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Union[str, Any] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : str = self.get_config() return config, pixel_values, labels def __snake_case ( self : List[str] ) -> Dict: return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act="gelu" , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=lowerCamelCase , layer_scale_init_value=1E-5 , ) def __snake_case ( self : int , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[int] , lowerCamelCase : List[str] ) -> List[Any]: __snake_case : Any = SwiftFormerModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Dict = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def __snake_case ( self : List[str] , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[int] ) -> List[Any]: __snake_case : Dict = self.num_labels __snake_case : Optional[Any] = SwiftFormerForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) __snake_case : List[Any] = SwiftFormerForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Any = model(lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Optional[int] ) -> List[str]: ((__snake_case) , (__snake_case) , (__snake_case)) : List[Any] = self.prepare_config_and_inputs() __snake_case : str = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Union[str, Any] = (SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () __UpperCAmelCase : Any = ( {"feature-extraction": SwiftFormerModel, "image-classification": SwiftFormerForImageClassification} if is_torch_available() else {} ) __UpperCAmelCase : Any = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : int = False __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : Dict = False def __snake_case ( self : Union[str, Any] ) -> Dict: __snake_case : int = SwiftFormerModelTester(self ) __snake_case : Union[str, Any] = ConfigTester( self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def __snake_case ( self : int ) -> Any: self.config_tester.run_common_tests() @unittest.skip(reason="SwiftFormer does not use inputs_embeds" ) def __snake_case ( self : str ) -> Tuple: pass def __snake_case ( self : Optional[Any] ) -> Tuple: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : List[str] = model_class(lowerCamelCase ) __snake_case : str = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase , nn.Linear ) ) def __snake_case ( self : Dict ) -> List[str]: __snake_case , __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : int = model_class(lowerCamelCase ) __snake_case : Any = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Optional[Any] = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> Union[str, Any]: __snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : int ) -> Any: __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def __snake_case ( self : Optional[int] ) -> Union[str, Any]: for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Dict = SwiftFormerModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @unittest.skip(reason="SwiftFormer does not output attentions" ) def __snake_case ( self : Tuple ) -> Union[str, Any]: pass def __snake_case ( self : Optional[int] ) -> List[Any]: def check_hidden_states_output(lowerCamelCase : Dict , lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict ): __snake_case : Any = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : Dict = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Tuple = outputs.hidden_states __snake_case : Optional[Any] = 8 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(lowerCamelCase ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) __snake_case , __snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Any = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : Union[str, Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Union[str, Any]: def _config_zero_init(lowerCamelCase : Optional[int] ): __snake_case : List[str] = copy.deepcopy(lowerCamelCase ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(lowerCamelCase , lowerCamelCase , 1E-10 ) if isinstance(getattr(lowerCamelCase , lowerCamelCase , lowerCamelCase ) , lowerCamelCase ): __snake_case : Optional[Any] = _config_zero_init(getattr(lowerCamelCase , lowerCamelCase ) ) setattr(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return configs_no_init __snake_case , __snake_case : Dict = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Dict = _config_zero_init(lowerCamelCase ) for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(config=lowerCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : List[Any] ) -> Any: pass def lowerCAmelCase_ ( ): __snake_case : Dict = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : int ) -> Any: return ViTImageProcessor.from_pretrained("MBZUAI/swiftformer-xs" ) if is_vision_available() else None @slow def __snake_case ( self : int ) -> Tuple: __snake_case : Dict = SwiftFormerForImageClassification.from_pretrained("MBZUAI/swiftformer-xs" ).to(lowerCamelCase ) __snake_case : Dict = self.default_image_processor __snake_case : int = prepare_img() __snake_case : Tuple = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : List[str] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : Tuple = torch.tensor([[-2.1703E00, 2.1107E00, -2.0811E00]] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) )
81
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) _snake_case : List[Any] = "\\n Text data.\n Second line of data." _snake_case : Tuple = "file" @pytest.fixture(scope="session" ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / (FILE_PATH + ".zstd") __snake_case : Optional[Any] = bytes(__lowerCamelCase , "utf-8" ) with zstd.open(__lowerCamelCase , "wb" ) as f: f.write(__lowerCamelCase ) return path @pytest.fixture def lowerCAmelCase_ ( __lowerCamelCase ): with open(os.path.join(tmpfs.local_root_dir , __lowerCamelCase ) , "w" ) as f: f.write(__lowerCamelCase ) return FILE_PATH @pytest.mark.parametrize("compression_format" , ["gzip", "xz", "zstd"] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Optional[int] = {"gzip": gz_file, "xz": xz_file, "zstd": zstd_path} __snake_case : str = input_paths[compression_format] __snake_case : Optional[Any] = tmp_path / "cache" __snake_case : Optional[int] = DownloadConfig(cache_dir=__lowerCamelCase , extract_compressed_file=__lowerCamelCase ) __snake_case : Union[str, Any] = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) with open(__lowerCamelCase ) as f: __snake_case : Dict = f.read() with open(__lowerCamelCase ) as f: __snake_case : Tuple = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize("default_extracted" , [True, False] ) @pytest.mark.parametrize("default_cache_dir" , [True, False] ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : Tuple = "custom_cache" __snake_case : List[str] = "custom_extracted_dir" __snake_case : Any = tmp_path / "custom_extracted_path" if default_extracted: __snake_case : List[Any] = ("downloads" if default_cache_dir else custom_cache_dir, "extracted") else: monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_DIR" , __lowerCamelCase ) monkeypatch.setattr("datasets.config.EXTRACTED_DATASETS_PATH" , str(__lowerCamelCase ) ) __snake_case : Optional[Any] = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) __snake_case : Optional[int] = xz_file __snake_case : Optional[int] = ( DownloadConfig(extract_compressed_file=__lowerCamelCase ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=__lowerCamelCase ) ) __snake_case : str = cached_path(__lowerCamelCase , download_config=__lowerCamelCase ) assert Path(__lowerCamelCase ).parent.parts[-2:] == expected def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : Optional[Any] = str(Path(__lowerCamelCase ).resolve() ) assert cached_path(__lowerCamelCase ) == text_file # relative path __snake_case : Any = str(Path(__lowerCamelCase ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(__lowerCamelCase ) == text_file def lowerCAmelCase_ ( __lowerCamelCase ): # absolute path __snake_case : List[Any] = str(tmp_path.resolve() / "__missing_file__.txt" ) with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) # relative path __snake_case : Optional[int] = "./__missing_file__.txt" with pytest.raises(__lowerCamelCase ): cached_path(__lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : str = get_from_cache(F'tmp://{tmpfs_file}' ) with open(__lowerCamelCase ) as f: __snake_case : Union[str, Any] = f.read() assert output_file_content == FILE_CONTENT @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( ): with pytest.raises(__lowerCamelCase ): cached_path("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[Any] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): http_get("https://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): http_head("https://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : List[str] = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): ftp_get("ftp://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): ftp_head("ftp://huggingface.co" ) @patch("datasets.config.HF_DATASETS_OFFLINE" , __lowerCamelCase ) def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Tuple = tmp_path_factory.mktemp("data" ) / "file.html" with pytest.raises(__lowerCamelCase ): fsspec_get("s3://huggingface.co" , temp_file=__lowerCamelCase ) with pytest.raises(__lowerCamelCase ): fsspec_head("s3://huggingface.co" )
81
1
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
_snake_case : Optional[int] = {"a": ["c", "b"], "b": ["d", "e"], "c": [], "d": [], "e": []} _snake_case : Dict = ["a", "b", "c", "d", "e"] def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __snake_case : List[str] = start # add current to visited visited.append(__lowerCamelCase ) __snake_case : List[Any] = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case : Tuple = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # if all neighbors visited add current to sort sort.append(__lowerCamelCase ) # if all vertices haven't been visited select a new one to visit if len(__lowerCamelCase ) != len(__lowerCamelCase ): for vertice in vertices: if vertice not in visited: __snake_case : int = topological_sort(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # return sort return sort if __name__ == "__main__": _snake_case : List[Any] = topological_sort("a", [], []) print(sort)
81
1
from __future__ import annotations import inspect import unittest import numpy as np from transformers import ResNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFResNetForImageClassification, TFResNetModel from transformers.models.resnet.modeling_tf_resnet import TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a : """simple docstring""" def __init__( self : int , lowerCamelCase : List[Any] , lowerCamelCase : List[Any]=3 , lowerCamelCase : Dict=32 , lowerCamelCase : List[Any]=3 , lowerCamelCase : List[str]=10 , lowerCamelCase : List[str]=[10, 20, 30, 40] , lowerCamelCase : Optional[Any]=[1, 1, 2, 1] , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Tuple=True , lowerCamelCase : int="relu" , lowerCamelCase : int=3 , lowerCamelCase : str=None , ) -> Optional[int]: __snake_case : str = parent __snake_case : List[Any] = batch_size __snake_case : int = image_size __snake_case : Optional[Any] = num_channels __snake_case : Any = embeddings_size __snake_case : Dict = hidden_sizes __snake_case : Tuple = depths __snake_case : List[str] = is_training __snake_case : int = use_labels __snake_case : str = hidden_act __snake_case : Tuple = num_labels __snake_case : Optional[int] = scope __snake_case : int = len(lowerCamelCase ) def __snake_case ( self : Tuple ) -> Optional[int]: __snake_case : int = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : str = None if self.use_labels: __snake_case : int = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = self.get_config() return config, pixel_values, labels def __snake_case ( self : List[Any] ) -> Optional[int]: return ResNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def __snake_case ( self : str , lowerCamelCase : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Optional[Any] ) -> Tuple: __snake_case : Dict = TFResNetModel(config=lowerCamelCase ) __snake_case : Dict = model(lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __snake_case ( self : List[Any] , lowerCamelCase : Tuple , lowerCamelCase : Union[str, Any] , lowerCamelCase : Optional[int] ) -> Optional[int]: __snake_case : Optional[int] = self.num_labels __snake_case : List[Any] = TFResNetForImageClassification(lowerCamelCase ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : int ) -> Dict: __snake_case : str = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case : Tuple = config_and_inputs __snake_case : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = (TFResNetModel, TFResNetForImageClassification) if is_tf_available() else () __UpperCAmelCase : Optional[int] = ( {"feature-extraction": TFResNetModel, "image-classification": TFResNetForImageClassification} if is_tf_available() else {} ) __UpperCAmelCase : Tuple = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : str = False __UpperCAmelCase : List[str] = False def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Any = TFResNetModelTester(self ) __snake_case : List[str] = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : List[Any] ) -> str: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __snake_case ( self : List[str] ) -> Optional[int]: return @unittest.skip(reason="ResNet does not use inputs_embeds" ) def __snake_case ( self : Union[str, Any] ) -> List[Any]: pass @unittest.skip(reason="ResNet does not support input and output embeddings" ) def __snake_case ( self : Tuple ) -> List[Any]: pass def __snake_case ( self : Optional[Any] ) -> Tuple: __snake_case , __snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[int] = model_class(lowerCamelCase ) __snake_case : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Tuple = [*signature.parameters.keys()] __snake_case : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : List[Any] ) -> str: __snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Tuple ) -> Optional[int]: def check_hidden_states_output(lowerCamelCase : int , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ): __snake_case : List[Any] = model_class(lowerCamelCase ) __snake_case : Tuple = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Optional[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __snake_case : Optional[int] = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase ) , expected_num_stages + 1 ) # ResNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __snake_case , __snake_case : str = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Optional[int] = ["basic", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: __snake_case : Tuple = layer_type __snake_case : Union[str, Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> int: __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Optional[int]: for model_name in TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : int = TFResNetModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Any = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : Dict ) -> str: return ( AutoImageProcessor.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __snake_case ( self : Optional[int] ) -> Tuple: __snake_case : Optional[int] = TFResNetForImageClassification.from_pretrained(TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) __snake_case : List[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : str = image_processor(images=lowerCamelCase , return_tensors="tf" ) # forward pass __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[str] = tf.constant([-11.10_69, -9.78_77, -8.37_77] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , lowerCamelCase , atol=1E-4 ) )
81
def lowerCAmelCase_ ( ): return [ a * b * (1_0_0_0 - a - b) for a in range(1 , 9_9_9 ) for b in range(__lowerCamelCase , 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(f'''{solution() = }''')
81
1
import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor _snake_case : List[str] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[str] , *lowerCamelCase : str , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DeiTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
from math import sqrt def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : int = 0 __snake_case : int = 0 __snake_case : int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(__lowerCamelCase , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f'''{solution() = }''')
81
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : Any , lowerCamelCase : str=7 , lowerCamelCase : List[str]=3 , lowerCamelCase : int=18 , lowerCamelCase : List[str]=30 , lowerCamelCase : Any=400 , lowerCamelCase : Optional[int]=True , lowerCamelCase : List[Any]=None , lowerCamelCase : Tuple=True , lowerCamelCase : Optional[int]=None , lowerCamelCase : Any=True , ) -> str: __snake_case : Dict = size if size is not None else {"shortest_edge": 20} __snake_case : Dict = crop_size if crop_size is not None else {"height": 18, "width": 18} __snake_case : Optional[Any] = parent __snake_case : Optional[int] = batch_size __snake_case : Optional[Any] = num_channels __snake_case : List[str] = image_size __snake_case : Optional[int] = min_resolution __snake_case : Optional[Any] = max_resolution __snake_case : List[Any] = do_resize __snake_case : List[Any] = size __snake_case : Optional[int] = do_center_crop __snake_case : int = crop_size __snake_case : List[str] = do_flip_channel_order def __snake_case ( self : Tuple ) -> List[str]: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Optional[int] = MobileViTImageProcessor if is_vision_available() else None def __snake_case ( self : Union[str, Any] ) -> str: __snake_case : int = MobileViTImageProcessingTester(self ) @property def __snake_case ( self : Optional[Any] ) -> Tuple: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Tuple ) -> Any: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "center_crop" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_flip_channel_order" ) ) def __snake_case ( self : List[Any] ) -> Union[str, Any]: __snake_case : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 20} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) __snake_case : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def __snake_case ( self : Optional[Any] ) -> Optional[Any]: pass def __snake_case ( self : Union[str, Any] ) -> int: # Initialize image_processing __snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : Tuple = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __snake_case ( self : Tuple ) -> List[Any]: # Initialize image_processing __snake_case : List[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : str = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Any = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched __snake_case : Tuple = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def __snake_case ( self : Dict ) -> int: # Initialize image_processing __snake_case : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : Optional[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched __snake_case : Any = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
81
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class a (_lowerCAmelCase ): """simple docstring""" def __snake_case ( self : str ) -> str: __snake_case : Tuple = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowerCamelCase , "hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "neck_hidden_sizes" ) ) self.parent.assertTrue(hasattr(lowerCamelCase , "num_attention_heads" ) ) class a : """simple docstring""" def __init__( self : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : Tuple=13 , lowerCamelCase : str=32 , lowerCamelCase : Dict=2 , lowerCamelCase : List[str]=3 , lowerCamelCase : Any=640 , lowerCamelCase : Optional[Any]=4 , lowerCamelCase : Tuple="silu" , lowerCamelCase : int=3 , lowerCamelCase : Dict=32 , lowerCamelCase : str=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Optional[Any]=0.1 , lowerCamelCase : Dict=0.02 , lowerCamelCase : Union[str, Any]=True , lowerCamelCase : Optional[int]=True , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : int=None , ) -> str: __snake_case : Optional[Any] = parent __snake_case : Optional[Any] = batch_size __snake_case : Any = image_size __snake_case : List[Any] = patch_size __snake_case : Any = num_channels __snake_case : Union[str, Any] = last_hidden_size __snake_case : Any = num_attention_heads __snake_case : Any = hidden_act __snake_case : Tuple = conv_kernel_size __snake_case : Any = output_stride __snake_case : Any = hidden_dropout_prob __snake_case : List[Any] = attention_probs_dropout_prob __snake_case : Optional[Any] = classifier_dropout_prob __snake_case : Union[str, Any] = use_labels __snake_case : Optional[int] = is_training __snake_case : Dict = num_labels __snake_case : Any = initializer_range __snake_case : Optional[int] = scope def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[Any] = None __snake_case : Optional[int] = None if self.use_labels: __snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Any = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels, pixel_labels def __snake_case ( self : Any ) -> Union[str, Any]: return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __snake_case ( self : Any , lowerCamelCase : Tuple , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Optional[int] ) -> Dict: __snake_case : List[Any] = MobileViTModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[str] = model(lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[str] , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple ) -> List[str]: __snake_case : str = self.num_labels __snake_case : List[Any] = MobileViTForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Dict , lowerCamelCase : Any , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict ) -> Dict: __snake_case : Union[str, Any] = self.num_labels __snake_case : Optional[int] = MobileViTForSemanticSegmentation(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Union[str, Any] = config_and_inputs __snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": MobileViTModel, "image-classification": MobileViTForImageClassification, "image-segmentation": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> Dict: __snake_case : Tuple = MobileViTModelTester(self ) __snake_case : Any = MobileViTConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase ) def __snake_case ( self : Optional[int] ) -> Dict: self.config_tester.run_common_tests() @unittest.skip(reason="MobileViT does not use inputs_embeds" ) def __snake_case ( self : Dict ) -> Any: pass @unittest.skip(reason="MobileViT does not support input and output embeddings" ) def __snake_case ( self : Dict ) -> List[Any]: pass @unittest.skip(reason="MobileViT does not output attentions" ) def __snake_case ( self : int ) -> Dict: pass def __snake_case ( self : int ) -> Union[str, Any]: __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = model_class(lowerCamelCase ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : int = [*signature.parameters.keys()] __snake_case : List[str] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def __snake_case ( self : int ) -> Tuple: pass def __snake_case ( self : Any ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: def check_hidden_states_output(lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : Any ): __snake_case : int = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : int = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.hidden_states __snake_case : int = 5 self.assertEqual(len(lowerCamelCase ) , lowerCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. __snake_case : List[Any] = 2 for i in range(len(lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) __snake_case , __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : List[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : Any ) -> Any: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> List[str]: __snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowerCamelCase ) @slow def __snake_case ( self : List[str] ) -> Any: for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileViTModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : str ) -> Dict: return MobileViTImageProcessor.from_pretrained("apple/mobilevit-xx-small" ) if is_vision_available() else None @slow def __snake_case ( self : Union[str, Any] ) -> List[str]: __snake_case : Tuple = MobileViTForImageClassification.from_pretrained("apple/mobilevit-xx-small" ).to(lowerCamelCase ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Union[str, Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Dict = model(**lowerCamelCase ) # verify the logits __snake_case : Union[str, Any] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : str ) -> Optional[int]: __snake_case : Optional[int] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : str = model.to(lowerCamelCase ) __snake_case : int = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Optional[int] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : List[str] = model(**lowerCamelCase ) __snake_case : Union[str, Any] = outputs.logits # verify the logits __snake_case : Tuple = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , lowerCamelCase ) __snake_case : Union[str, Any] = torch.tensor( [ [[6.97_13, 6.97_86, 7.24_22], [7.28_93, 7.28_25, 7.44_46], [7.65_80, 7.87_97, 7.94_20]], [[-10.68_69, -10.32_50, -10.34_71], [-10.42_28, -9.98_68, -9.71_32], [-11.04_05, -11.02_21, -10.73_18]], [[-3.30_89, -2.85_39, -2.67_40], [-3.27_06, -2.56_21, -2.51_08], [-3.25_34, -2.66_15, -2.66_51]], ] , device=lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowerCamelCase , atol=1E-4 ) ) @slow def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = MobileViTForSemanticSegmentation.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : Tuple = model.to(lowerCamelCase ) __snake_case : Dict = MobileViTImageProcessor.from_pretrained("apple/deeplabv3-mobilevit-xx-small" ) __snake_case : List[Any] = prepare_img() __snake_case : List[Any] = image_processor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : Any = model(**lowerCamelCase ) __snake_case : Dict = outputs.logits.detach().cpu() __snake_case : Any = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase , target_sizes=[(50, 60)] ) __snake_case : int = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , lowerCamelCase ) __snake_case : List[str] = image_processor.post_process_semantic_segmentation(outputs=lowerCamelCase ) __snake_case : Optional[int] = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , lowerCamelCase )
81
1
from __future__ import annotations _snake_case : Union[str, Any] = { "A": ["B", "C", "E"], "B": ["A", "D", "E"], "C": ["A", "F", "G"], "D": ["B"], "E": ["A", "B", "D"], "F": ["C"], "G": ["C"], } class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : dict[str, list[str]] , lowerCamelCase : str ) -> None: __snake_case : Optional[int] = graph # mapping node to its parent in resulting breadth first tree __snake_case : dict[str, str | None] = {} __snake_case : Optional[Any] = source_vertex def __snake_case ( self : Tuple ) -> None: __snake_case : Optional[int] = {self.source_vertex} __snake_case : Union[str, Any] = None __snake_case : List[str] = [self.source_vertex] # first in first out queue while queue: __snake_case : Union[str, Any] = queue.pop(0 ) for adjacent_vertex in self.graph[vertex]: if adjacent_vertex not in visited: visited.add(lowerCamelCase ) __snake_case : Tuple = vertex queue.append(lowerCamelCase ) def __snake_case ( self : Optional[int] , lowerCamelCase : str ) -> str: if target_vertex == self.source_vertex: return self.source_vertex __snake_case : List[Any] = self.parent.get(lowerCamelCase ) if target_vertex_parent is None: __snake_case : str = ( F'No path from vertex: {self.source_vertex} to vertex: {target_vertex}' ) raise ValueError(lowerCamelCase ) return self.shortest_path(lowerCamelCase ) + F'->{target_vertex}' if __name__ == "__main__": _snake_case : Optional[int] = Graph(graph, "G") g.breath_first_search() print(g.shortest_path("D")) print(g.shortest_path("G")) print(g.shortest_path("Foo"))
81
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor _snake_case : int = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[int] , *lowerCamelCase : List[Any] , **lowerCamelCase : int ) -> None: warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
1
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = AutoencoderKL __UpperCAmelCase : Optional[Any] = "sample" __UpperCAmelCase : Optional[int] = 1e-2 @property def __snake_case ( self : Dict ) -> Optional[Any]: __snake_case : Optional[Any] = 4 __snake_case : Tuple = 3 __snake_case : List[str] = (32, 32) __snake_case : str = floats_tensor((batch_size, num_channels) + sizes ).to(lowerCamelCase ) return {"sample": image} @property def __snake_case ( self : Union[str, Any] ) -> Tuple: return (3, 32, 32) @property def __snake_case ( self : int ) -> int: return (3, 32, 32) def __snake_case ( self : Optional[Any] ) -> Dict: __snake_case : Optional[Any] = { "block_out_channels": [32, 64], "in_channels": 3, "out_channels": 3, "down_block_types": ["DownEncoderBlock2D", "DownEncoderBlock2D"], "up_block_types": ["UpDecoderBlock2D", "UpDecoderBlock2D"], "latent_channels": 4, } __snake_case : Any = self.dummy_input return init_dict, inputs_dict def __snake_case ( self : str ) -> Dict: pass def __snake_case ( self : Tuple ) -> List[str]: pass @unittest.skipIf(torch_device == "mps" , "Gradient checkpointing skipped on MPS" ) def __snake_case ( self : Any ) -> Optional[Any]: # enable deterministic behavior for gradient checkpointing __snake_case , __snake_case : int = self.prepare_init_args_and_inputs_for_common() __snake_case : str = self.model_class(**lowerCamelCase ) model.to(lowerCamelCase ) assert not model.is_gradient_checkpointing and model.training __snake_case : str = model(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model.zero_grad() __snake_case : Any = torch.randn_like(lowerCamelCase ) __snake_case : str = (out - labels).mean() loss.backward() # re-instantiate the model now enabling gradient checkpointing __snake_case : Optional[int] = self.model_class(**lowerCamelCase ) # clone model model_a.load_state_dict(model.state_dict() ) model_a.to(lowerCamelCase ) model_a.enable_gradient_checkpointing() assert model_a.is_gradient_checkpointing and model_a.training __snake_case : int = model_a(**lowerCamelCase ).sample # run the backwards pass on the model. For backwards pass, for simplicity purpose, # we won't calculate the loss and rather backprop on out.sum() model_a.zero_grad() __snake_case : Union[str, Any] = (out_a - labels).mean() loss_a.backward() # compare the output and parameters gradients self.assertTrue((loss - loss_a).abs() < 1E-5 ) __snake_case : Optional[int] = dict(model.named_parameters() ) __snake_case : List[Any] = dict(model_a.named_parameters() ) for name, param in named_params.items(): self.assertTrue(torch_all_close(param.grad.data , named_params_a[name].grad.data , atol=5E-5 ) ) def __snake_case ( self : List[Any] ) -> Optional[int]: __snake_case , __snake_case : Optional[Any] = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" , output_loading_info=lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) self.assertEqual(len(loading_info["missing_keys"] ) , 0 ) model.to(lowerCamelCase ) __snake_case : Optional[Any] = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def __snake_case ( self : Optional[Any] ) -> Union[str, Any]: __snake_case : Tuple = AutoencoderKL.from_pretrained("fusing/autoencoder-kl-dummy" ) __snake_case : Dict = model.to(lowerCamelCase ) model.eval() if torch_device == "mps": __snake_case : int = torch.manual_seed(0 ) else: __snake_case : str = torch.Generator(device=lowerCamelCase ).manual_seed(0 ) __snake_case : List[str] = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) __snake_case : Union[str, Any] = image.to(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , sample_posterior=lowerCamelCase , generator=lowerCamelCase ).sample __snake_case : List[Any] = output[0, -1, -3:, -3:].flatten().cpu() # Since the VAE Gaussian prior's generator is seeded on the appropriate device, # the expected output slices are not the same for CPU and GPU. if torch_device == "mps": __snake_case : Union[str, Any] = torch.tensor( [ -4.0078E-01, -3.8323E-04, -1.2681E-01, -1.1462E-01, 2.0095E-01, 1.0893E-01, -8.8247E-02, -3.0361E-01, -9.8644E-03, ] ) elif torch_device == "cpu": __snake_case : Tuple = torch.tensor( [-0.13_52, 0.08_78, 0.04_19, -0.08_18, -0.10_69, 0.06_88, -0.14_58, -0.44_46, -0.00_26] ) else: __snake_case : List[str] = torch.tensor( [-0.24_21, 0.46_42, 0.25_07, -0.04_38, 0.06_82, 0.31_60, -0.20_18, -0.07_27, 0.24_85] ) self.assertTrue(torch_all_close(lowerCamelCase , lowerCamelCase , rtol=1E-2 ) ) @slow class a (unittest.TestCase ): """simple docstring""" def __snake_case ( self : int , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] ) -> List[str]: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase ) for s in shape] )}.npy' def __snake_case ( self : List[Any] ) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def __snake_case ( self : Tuple , lowerCamelCase : List[Any]=0 , lowerCamelCase : Tuple=(4, 3, 512, 512) , lowerCamelCase : Optional[int]=False ) -> str: __snake_case : List[Any] = torch.floataa if fpaa else torch.floataa __snake_case : Tuple = torch.from_numpy(load_hf_numpy(self.get_file_format(lowerCamelCase , lowerCamelCase ) ) ).to(lowerCamelCase ).to(lowerCamelCase ) return image def __snake_case ( self : Optional[Any] , lowerCamelCase : int="CompVis/stable-diffusion-v1-4" , lowerCamelCase : int=False ) -> int: __snake_case : str = "fp16" if fpaa else None __snake_case : int = torch.floataa if fpaa else torch.floataa __snake_case : int = AutoencoderKL.from_pretrained( lowerCamelCase , subfolder="vae" , torch_dtype=lowerCamelCase , revision=lowerCamelCase , ) model.to(lowerCamelCase ).eval() return model def __snake_case ( self : str , lowerCamelCase : int=0 ) -> Optional[Any]: if torch_device == "mps": return torch.manual_seed(lowerCamelCase ) return torch.Generator(device=lowerCamelCase ).manual_seed(lowerCamelCase ) @parameterized.expand( [ # fmt: off [33, [-0.16_03, 0.98_78, -0.04_95, -0.07_90, -0.27_09, 0.83_75, -0.20_60, -0.08_24], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_76, 0.11_68, 0.13_32, -0.48_40, -0.25_08, -0.07_91, -0.04_93, -0.40_89], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[str] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] ) -> List[Any]: __snake_case : Optional[Any] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) __snake_case : Tuple = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : List[Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : int = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [33, [-0.05_13, 0.02_89, 1.37_99, 0.21_66, -0.25_73, -0.08_71, 0.51_03, -0.09_99]], [47, [-0.41_28, -0.13_20, -0.37_04, 0.19_65, -0.41_16, -0.23_32, -0.33_40, 0.22_47]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : Any , lowerCamelCase : List[str] , lowerCamelCase : List[str] ) -> Tuple: __snake_case : Any = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , fpaa=lowerCamelCase ) __snake_case : List[Any] = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : str = model(lowerCamelCase , generator=lowerCamelCase , sample_posterior=lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Any = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.16_09, 0.98_66, -0.04_87, -0.07_77, -0.27_16, 0.83_68, -0.20_55, -0.08_14], [-0.23_95, 0.00_98, 0.01_02, -0.07_09, -0.28_40, -0.02_74, -0.07_18, -0.18_24]], [47, [-0.23_77, 0.11_47, 0.13_33, -0.48_41, -0.25_06, -0.08_05, -0.04_91, -0.40_85], [0.03_50, 0.08_47, 0.04_67, 0.03_44, -0.08_42, -0.05_47, -0.06_33, -0.11_31]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase ) with torch.no_grad(): __snake_case : int = model(lowerCamelCase ).sample assert sample.shape == image.shape __snake_case : Union[str, Any] = sample[-1, -2:, -2:, :2].flatten().float().cpu() __snake_case : List[str] = torch.tensor(expected_slice_mps if torch_device == "mps" else expected_slice ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=3E-3 ) @parameterized.expand( [ # fmt: off [13, [-0.20_51, -0.18_03, -0.23_11, -0.21_14, -0.32_92, -0.35_74, -0.29_53, -0.33_23]], [37, [-0.26_32, -0.26_25, -0.21_99, -0.27_41, -0.45_39, -0.49_90, -0.37_20, -0.49_25]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : List[str] , lowerCamelCase : Tuple , lowerCamelCase : Any ) -> Optional[Any]: __snake_case : List[str] = self.get_sd_vae_model() __snake_case : List[Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : str = sample[-1, -2:, :2, -2:].flatten().cpu() __snake_case : Optional[int] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-3 ) @parameterized.expand( [ # fmt: off [27, [-0.03_69, 0.02_07, -0.07_76, -0.06_82, -0.17_47, -0.19_30, -0.14_65, -0.20_39]], [16, [-0.16_28, -0.21_34, -0.27_47, -0.26_42, -0.37_74, -0.44_04, -0.36_87, -0.42_77]], # fmt: on ] ) @require_torch_gpu def __snake_case ( self : str , lowerCamelCase : Optional[int] , lowerCamelCase : Dict ) -> int: __snake_case : int = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : List[str] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : Union[str, Any] = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] __snake_case : Optional[Any] = sample[-1, -2:, :2, -2:].flatten().float().cpu() __snake_case : Optional[Any] = torch.tensor(lowerCamelCase ) assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=5E-3 ) @parameterized.expand([(13,), (16,), (27,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> Tuple: __snake_case : Dict = self.get_sd_vae_model(fpaa=lowerCamelCase ) __snake_case : Any = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) , fpaa=lowerCamelCase ) with torch.no_grad(): __snake_case : str = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Any = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-1 ) @parameterized.expand([(13,), (16,), (37,)] ) @require_torch_gpu @unittest.skipIf(not is_xformers_available() , reason="xformers is not required when using PyTorch 2.0." ) def __snake_case ( self : List[Any] , lowerCamelCase : Any ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : Union[str, Any] = self.get_sd_image(lowerCamelCase , shape=(3, 4, 64, 64) ) with torch.no_grad(): __snake_case : List[Any] = model.decode(lowerCamelCase ).sample model.enable_xformers_memory_efficient_attention() with torch.no_grad(): __snake_case : Dict = model.decode(lowerCamelCase ).sample assert list(sample.shape ) == [3, 3, 512, 512] assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=1E-2 ) @parameterized.expand( [ # fmt: off [33, [-0.30_01, 0.09_18, -2.69_84, -3.97_20, -3.20_99, -5.03_53, 1.73_38, -0.20_65, 3.42_67]], [47, [-1.50_30, -4.38_71, -6.03_55, -9.11_57, -1.66_61, -2.78_53, 2.16_07, -5.08_23, 2.56_33]], # fmt: on ] ) def __snake_case ( self : List[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Dict ) -> Optional[int]: __snake_case : str = self.get_sd_vae_model() __snake_case : int = self.get_sd_image(lowerCamelCase ) __snake_case : int = self.get_generator(lowerCamelCase ) with torch.no_grad(): __snake_case : Optional[Any] = model.encode(lowerCamelCase ).latent_dist __snake_case : Dict = dist.sample(generator=lowerCamelCase ) assert list(sample.shape ) == [image.shape[0], 4] + [i // 8 for i in image.shape[2:]] __snake_case : List[str] = sample[0, -1, -3:, -3:].flatten().cpu() __snake_case : Dict = torch.tensor(lowerCamelCase ) __snake_case : Dict = 3E-3 if torch_device != "mps" else 1E-2 assert torch_all_close(lowerCamelCase , lowerCamelCase , atol=lowerCamelCase )
81
from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging _snake_case : Optional[Any] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : List[str] = ["pixel_values"] def __init__( self : List[Any] , lowerCamelCase : bool = True , lowerCamelCase : Union[int, float] = 1 / 255 , lowerCamelCase : bool = True , lowerCamelCase : int = 8 , **lowerCamelCase : Tuple , ) -> None: super().__init__(**lowerCamelCase ) __snake_case : Dict = do_rescale __snake_case : Dict = rescale_factor __snake_case : Optional[Any] = do_pad __snake_case : Tuple = pad_size def __snake_case ( self : Dict , lowerCamelCase : np.ndarray , lowerCamelCase : float , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase : Optional[int] ) -> np.ndarray: return rescale(lowerCamelCase , scale=lowerCamelCase , data_format=lowerCamelCase , **lowerCamelCase ) def __snake_case ( self : Optional[Any] , lowerCamelCase : np.ndarray , lowerCamelCase : int , lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ) -> Tuple: __snake_case , __snake_case : List[str] = get_image_size(lowerCamelCase ) __snake_case : Optional[Any] = (old_height // size + 1) * size - old_height __snake_case : List[Any] = (old_width // size + 1) * size - old_width return pad(lowerCamelCase , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=lowerCamelCase ) def __snake_case ( self : Tuple , lowerCamelCase : ImageInput , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[float] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Union[str, ChannelDimension] = ChannelDimension.FIRST , **lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : int = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : str = do_pad if do_pad is not None else self.do_pad __snake_case : Any = pad_size if pad_size is not None else self.pad_size __snake_case : int = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case : str = [to_numpy_array(lowerCamelCase ) for image in images] if do_rescale: __snake_case : Optional[int] = [self.rescale(image=lowerCamelCase , scale=lowerCamelCase ) for image in images] if do_pad: __snake_case : Optional[Any] = [self.pad(lowerCamelCase , size=lowerCamelCase ) for image in images] __snake_case : int = [to_channel_dimension_format(lowerCamelCase , lowerCamelCase ) for image in images] __snake_case : Union[str, Any] = {"pixel_values": images} return BatchFeature(data=lowerCamelCase , tensor_type=lowerCamelCase )
81
1
import qiskit def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase ): __snake_case : List[Any] = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register __snake_case : Optional[Any] = qiskit.QuantumCircuit(__lowerCamelCase , __lowerCamelCase ) # Map the quantum measurement to the classical bits circuit.measure([0] , [0] ) # Execute the circuit on the simulator __snake_case : List[Any] = qiskit.execute(__lowerCamelCase , __lowerCamelCase , shots=1_0_0_0 ) # Return the histogram data of the results of the experiment. return job.result().get_counts(__lowerCamelCase ) if __name__ == "__main__": print(f'''Total count for various states are: {single_qubit_measure(1, 1)}''')
81
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case : Any = { "configuration_informer": [ "INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "InformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ "INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "InformerForPrediction", "InformerModel", "InformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
1
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple , lowerCamelCase : Optional[int]=13 , lowerCamelCase : str=7 , lowerCamelCase : Any=True , lowerCamelCase : Tuple=True , lowerCamelCase : int=True , lowerCamelCase : int=True , lowerCamelCase : Dict=True , lowerCamelCase : Tuple=False , lowerCamelCase : Union[str, Any]=False , lowerCamelCase : Union[str, Any]=False , lowerCamelCase : int=2 , lowerCamelCase : Optional[int]=99 , lowerCamelCase : List[Any]=0 , lowerCamelCase : Any=32 , lowerCamelCase : Optional[Any]=5 , lowerCamelCase : int=4 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : List[Any]=0.1 , lowerCamelCase : Tuple=512 , lowerCamelCase : Optional[Any]=12 , lowerCamelCase : Optional[int]=2 , lowerCamelCase : Optional[int]=0.02 , lowerCamelCase : Union[str, Any]=3 , lowerCamelCase : List[str]=4 , lowerCamelCase : Any="last" , lowerCamelCase : Optional[int]=None , lowerCamelCase : Optional[Any]=None , ) -> Dict: __snake_case : Any = parent __snake_case : Any = batch_size __snake_case : Optional[Any] = seq_length __snake_case : Optional[int] = is_training __snake_case : List[Any] = use_input_lengths __snake_case : Optional[Any] = use_token_type_ids __snake_case : str = use_labels __snake_case : Union[str, Any] = gelu_activation __snake_case : Optional[int] = sinusoidal_embeddings __snake_case : Optional[int] = causal __snake_case : List[str] = asm __snake_case : Optional[int] = n_langs __snake_case : Tuple = vocab_size __snake_case : Optional[int] = n_special __snake_case : Any = hidden_size __snake_case : List[Any] = num_hidden_layers __snake_case : List[str] = num_attention_heads __snake_case : Any = hidden_dropout_prob __snake_case : Dict = attention_probs_dropout_prob __snake_case : Optional[Any] = max_position_embeddings __snake_case : Union[str, Any] = type_vocab_size __snake_case : List[str] = type_sequence_label_size __snake_case : List[str] = initializer_range __snake_case : Optional[Any] = num_labels __snake_case : List[str] = num_choices __snake_case : Optional[Any] = summary_type __snake_case : int = use_proj __snake_case : Optional[Any] = scope def __snake_case ( self : Any ) -> List[Any]: __snake_case : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case : int = None if self.use_input_lengths: __snake_case : Any = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length __snake_case : Any = None if self.use_token_type_ids: __snake_case : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) __snake_case : str = None __snake_case : Optional[Any] = None __snake_case : List[str] = None if self.use_labels: __snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case : Tuple = ids_tensor([self.batch_size] , 2 ).float() __snake_case : Any = ids_tensor([self.batch_size] , self.num_choices ) __snake_case : List[str] = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def __snake_case ( self : Optional[Any] ) -> Dict: return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def __snake_case ( self : Tuple , lowerCamelCase : Optional[int] , lowerCamelCase : str , lowerCamelCase : str , lowerCamelCase : Union[str, Any] , lowerCamelCase : Tuple , lowerCamelCase : Any , lowerCamelCase : str , lowerCamelCase : Dict , lowerCamelCase : Optional[Any] , ) -> Dict: __snake_case : List[str] = FlaubertModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : str = model(lowerCamelCase , lengths=lowerCamelCase , langs=lowerCamelCase ) __snake_case : List[Any] = model(lowerCamelCase , langs=lowerCamelCase ) __snake_case : Tuple = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __snake_case ( self : List[str] , lowerCamelCase : Union[str, Any] , lowerCamelCase : str , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict , lowerCamelCase : List[str] , lowerCamelCase : List[Any] , lowerCamelCase : Tuple , ) -> Optional[Any]: __snake_case : Optional[Any] = FlaubertWithLMHeadModel(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Dict = model(lowerCamelCase , token_type_ids=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case ( self : Dict , lowerCamelCase : List[str] , lowerCamelCase : Optional[int] , lowerCamelCase : Optional[int] , lowerCamelCase : List[str] , lowerCamelCase : List[str] , lowerCamelCase : int , lowerCamelCase : Tuple , lowerCamelCase : Dict , lowerCamelCase : List[Any] , ) -> List[str]: __snake_case : Union[str, Any] = FlaubertForQuestionAnsweringSimple(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Optional[int] = model(lowerCamelCase ) __snake_case : Optional[int] = model(lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __snake_case ( self : Tuple , lowerCamelCase : Optional[int] , lowerCamelCase : int , lowerCamelCase : Union[str, Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple , lowerCamelCase : str , lowerCamelCase : Tuple , lowerCamelCase : Dict , ) -> Union[str, Any]: __snake_case : str = FlaubertForQuestionAnswering(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : int = model(lowerCamelCase ) __snake_case : Dict = model( lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase , cls_index=lowerCamelCase , is_impossible=lowerCamelCase , p_mask=lowerCamelCase , ) __snake_case : List[Any] = model( lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase , cls_index=lowerCamelCase , is_impossible=lowerCamelCase , ) ((__snake_case) , ) : Union[str, Any] = result_with_labels.to_tuple() __snake_case : List[Any] = model(lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase ) ((__snake_case) , ) : List[str] = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : int , lowerCamelCase : str , lowerCamelCase : Any , lowerCamelCase : int , lowerCamelCase : List[str] , lowerCamelCase : str , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple , lowerCamelCase : List[Any] , ) -> str: __snake_case : Union[str, Any] = FlaubertForSequenceClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Dict = model(lowerCamelCase ) __snake_case : List[Any] = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __snake_case ( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : Optional[Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Dict , lowerCamelCase : Optional[int] , lowerCamelCase : Union[str, Any] , ) -> str: __snake_case : Any = self.num_labels __snake_case : Union[str, Any] = FlaubertForTokenClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : int = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __snake_case ( self : Optional[Any] , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Optional[Any] , lowerCamelCase : Union[str, Any] , lowerCamelCase : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any] , lowerCamelCase : Union[str, Any] , ) -> List[str]: __snake_case : List[str] = self.num_choices __snake_case : Any = FlaubertForMultipleChoice(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Any = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case : Tuple = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case : str = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case : Union[str, Any] = model( lowerCamelCase , attention_mask=lowerCamelCase , token_type_ids=lowerCamelCase , labels=lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __snake_case ( self : Tuple ) -> Tuple: __snake_case : int = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) : int = config_and_inputs __snake_case : str = { "input_ids": input_ids, "token_type_ids": token_type_ids, "lengths": input_lengths, "attention_mask": input_mask, } return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Dict = ( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) __UpperCAmelCase : Optional[Any] = ( { "feature-extraction": FlaubertModel, "fill-mask": FlaubertWithLMHeadModel, "question-answering": FlaubertForQuestionAnsweringSimple, "text-classification": FlaubertForSequenceClassification, "token-classification": FlaubertForTokenClassification, "zero-shot": FlaubertForSequenceClassification, } if is_torch_available() else {} ) def __snake_case ( self : List[str] , lowerCamelCase : int , lowerCamelCase : Dict , lowerCamelCase : List[str] , lowerCamelCase : int , lowerCamelCase : Union[str, Any] ) -> Tuple: if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("Fast" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def __snake_case ( self : Optional[int] , lowerCamelCase : int , lowerCamelCase : Optional[Any] , lowerCamelCase : List[str]=False ) -> List[str]: __snake_case : Optional[Any] = super()._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": __snake_case : Optional[int] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase ) __snake_case : Optional[Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase ) return inputs_dict def __snake_case ( self : Tuple ) -> int: __snake_case : int = FlaubertModelTester(self ) __snake_case : Union[str, Any] = ConfigTester(self , config_class=lowerCamelCase , emb_dim=37 ) def __snake_case ( self : str ) -> Dict: self.config_tester.run_common_tests() def __snake_case ( self : Union[str, Any] ) -> Optional[int]: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*lowerCamelCase ) def __snake_case ( self : Any ) -> str: __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*lowerCamelCase ) def __snake_case ( self : Any ) -> Union[str, Any]: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*lowerCamelCase ) def __snake_case ( self : List[str] ) -> Optional[int]: __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*lowerCamelCase ) def __snake_case ( self : int ) -> Tuple: __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*lowerCamelCase ) def __snake_case ( self : int ) -> Tuple: __snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*lowerCamelCase ) def __snake_case ( self : Union[str, Any] ) -> int: __snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*lowerCamelCase ) @slow def __snake_case ( self : Any ) -> str: for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : Dict = FlaubertModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @slow @require_torch_gpu def __snake_case ( self : str ) -> int: __snake_case , __snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return __snake_case : str = True __snake_case : List[Any] = model_class(config=lowerCamelCase ) __snake_case : Union[str, Any] = self._prepare_for_class(lowerCamelCase , lowerCamelCase ) __snake_case : str = torch.jit.trace( lowerCamelCase , (inputs_dict["input_ids"].to("cpu" ), inputs_dict["attention_mask"].to("cpu" )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowerCamelCase , os.path.join(lowerCamelCase , "traced_model.pt" ) ) __snake_case : Union[str, Any] = torch.jit.load(os.path.join(lowerCamelCase , "traced_model.pt" ) , map_location=lowerCamelCase ) loaded(inputs_dict["input_ids"].to(lowerCamelCase ) , inputs_dict["attention_mask"].to(lowerCamelCase ) ) @require_torch class a (unittest.TestCase ): """simple docstring""" @slow def __snake_case ( self : str ) -> List[Any]: __snake_case : int = FlaubertModel.from_pretrained("flaubert/flaubert_base_cased" ) __snake_case : Dict = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) with torch.no_grad(): __snake_case : List[str] = model(lowerCamelCase )[0] __snake_case : Any = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , lowerCamelCase ) __snake_case : Dict = torch.tensor( [[[-2.62_51, -1.42_98, -0.02_27], [-2.85_10, -1.63_87, 0.22_58], [-2.81_14, -1.18_32, -0.30_66]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase , atol=1E-4 ) )
81
from typing import Callable, List, Optional, Union import PIL import torch from transformers import ( CLIPImageProcessor, CLIPSegForImageSegmentation, CLIPSegProcessor, CLIPTextModel, CLIPTokenizer, ) from diffusers import DiffusionPipeline from diffusers.configuration_utils import FrozenDict from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stable_diffusion import StableDiffusionInpaintPipeline from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.schedulers import DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler from diffusers.utils import deprecate, is_accelerate_available, logging _snake_case : Any = logging.get_logger(__name__) # pylint: disable=invalid-name class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : List[Any] , lowerCamelCase : CLIPSegForImageSegmentation , lowerCamelCase : CLIPSegProcessor , lowerCamelCase : AutoencoderKL , lowerCamelCase : CLIPTextModel , lowerCamelCase : CLIPTokenizer , lowerCamelCase : UNetaDConditionModel , lowerCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , lowerCamelCase : StableDiffusionSafetyChecker , lowerCamelCase : CLIPImageProcessor , ) -> Tuple: super().__init__() if hasattr(scheduler.config , "steps_offset" ) and scheduler.config.steps_offset != 1: __snake_case : Tuple = ( F'The configuration file of this scheduler: {scheduler} is outdated. `steps_offset`' F' should be set to 1 instead of {scheduler.config.steps_offset}. Please make sure ' "to update the config accordingly as leaving `steps_offset` might led to incorrect results" " in future versions. If you have downloaded this checkpoint from the Hugging Face Hub," " it would be very nice if you could open a Pull request for the `scheduler/scheduler_config.json`" " file" ) deprecate("steps_offset!=1" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : Any = dict(scheduler.config ) __snake_case : List[Any] = 1 __snake_case : Tuple = FrozenDict(lowerCamelCase ) if hasattr(scheduler.config , "skip_prk_steps" ) and scheduler.config.skip_prk_steps is False: __snake_case : List[str] = ( F'The configuration file of this scheduler: {scheduler} has not set the configuration' " `skip_prk_steps`. `skip_prk_steps` should be set to True in the configuration file. Please make" " sure to update the config accordingly as not setting `skip_prk_steps` in the config might lead to" " incorrect results in future versions. If you have downloaded this checkpoint from the Hugging Face" " Hub, it would be very nice if you could open a Pull request for the" " `scheduler/scheduler_config.json` file" ) deprecate("skip_prk_steps not set" , "1.0.0" , lowerCamelCase , standard_warn=lowerCamelCase ) __snake_case : List[str] = dict(scheduler.config ) __snake_case : List[str] = True __snake_case : Any = FrozenDict(lowerCamelCase ) if safety_checker is None: logger.warning( F'You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( segmentation_model=lowerCamelCase , segmentation_processor=lowerCamelCase , vae=lowerCamelCase , text_encoder=lowerCamelCase , tokenizer=lowerCamelCase , unet=lowerCamelCase , scheduler=lowerCamelCase , safety_checker=lowerCamelCase , feature_extractor=lowerCamelCase , ) def __snake_case ( self : Dict , lowerCamelCase : Optional[Union[str, int]] = "auto" ) -> Tuple: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory __snake_case : Any = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase ) def __snake_case ( self : List[Any] ) -> Any: self.enable_attention_slicing(lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) __snake_case : Optional[int] = torch.device("cuda" ) for cpu_offloaded_model in [self.unet, self.text_encoder, self.vae, self.safety_checker]: if cpu_offloaded_model is not None: cpu_offload(lowerCamelCase , lowerCamelCase ) @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __snake_case ( self : int ) -> Any: if self.device != torch.device("meta" ) or not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowerCamelCase , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() def __call__( self : List[Any] , lowerCamelCase : Union[str, List[str]] , lowerCamelCase : Union[torch.FloatTensor, PIL.Image.Image] , lowerCamelCase : str , lowerCamelCase : int = 512 , lowerCamelCase : int = 512 , lowerCamelCase : int = 50 , lowerCamelCase : float = 7.5 , lowerCamelCase : Optional[Union[str, List[str]]] = None , lowerCamelCase : Optional[int] = 1 , lowerCamelCase : float = 0.0 , lowerCamelCase : Optional[torch.Generator] = None , lowerCamelCase : Optional[torch.FloatTensor] = None , lowerCamelCase : Optional[str] = "pil" , lowerCamelCase : bool = True , lowerCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , lowerCamelCase : int = 1 , **lowerCamelCase : Dict , ) -> List[str]: __snake_case : Tuple = self.segmentation_processor( text=[text] , images=[image] , padding="max_length" , return_tensors="pt" ).to(self.device ) __snake_case : str = self.segmentation_model(**lowerCamelCase ) __snake_case : Dict = torch.sigmoid(outputs.logits ).cpu().detach().unsqueeze(-1 ).numpy() __snake_case : List[Any] = self.numpy_to_pil(lowerCamelCase )[0].resize(image.size ) # Run inpainting pipeline with the generated mask __snake_case : Tuple = StableDiffusionInpaintPipeline( vae=self.vae , text_encoder=self.text_encoder , tokenizer=self.tokenizer , unet=self.unet , scheduler=self.scheduler , safety_checker=self.safety_checker , feature_extractor=self.feature_extractor , ) return inpainting_pipeline( prompt=lowerCamelCase , image=lowerCamelCase , mask_image=lowerCamelCase , height=lowerCamelCase , width=lowerCamelCase , num_inference_steps=lowerCamelCase , guidance_scale=lowerCamelCase , negative_prompt=lowerCamelCase , num_images_per_prompt=lowerCamelCase , eta=lowerCamelCase , generator=lowerCamelCase , latents=lowerCamelCase , output_type=lowerCamelCase , return_dict=lowerCamelCase , callback=lowerCamelCase , callback_steps=lowerCamelCase , )
81
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _snake_case : str = { "configuration_whisper": ["WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP", "WhisperConfig", "WhisperOnnxConfig"], "feature_extraction_whisper": ["WhisperFeatureExtractor"], "processing_whisper": ["WhisperProcessor"], "tokenization_whisper": ["WhisperTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = ["WhisperTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : List[str] = [ "WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "WhisperForConditionalGeneration", "WhisperModel", "WhisperPreTrainedModel", "WhisperForAudioClassification", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[int] = [ "TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFWhisperForConditionalGeneration", "TFWhisperModel", "TFWhisperPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : Optional[Any] = [ "FlaxWhisperForConditionalGeneration", "FlaxWhisperModel", "FlaxWhisperPreTrainedModel", "FlaxWhisperForAudioClassification", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _snake_case : Optional[int] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
81
class a : """simple docstring""" def __init__( self : Tuple , lowerCamelCase : list ) -> None: __snake_case : str = set_counts __snake_case : Union[str, Any] = max(lowerCamelCase ) __snake_case : List[Any] = len(lowerCamelCase ) __snake_case : Tuple = [1] * num_sets __snake_case : Dict = list(range(lowerCamelCase ) ) def __snake_case ( self : str , lowerCamelCase : int , lowerCamelCase : int ) -> bool: __snake_case : List[Any] = self.get_parent(lowerCamelCase ) __snake_case : Tuple = self.get_parent(lowerCamelCase ) if src_parent == dst_parent: return False if self.ranks[dst_parent] >= self.ranks[src_parent]: self.set_counts[dst_parent] += self.set_counts[src_parent] __snake_case : List[str] = 0 __snake_case : List[Any] = dst_parent if self.ranks[dst_parent] == self.ranks[src_parent]: self.ranks[dst_parent] += 1 __snake_case : Dict = self.set_counts[dst_parent] else: self.set_counts[src_parent] += self.set_counts[dst_parent] __snake_case : Union[str, Any] = 0 __snake_case : Optional[int] = src_parent __snake_case : Tuple = self.set_counts[src_parent] __snake_case : str = max(self.max_set , lowerCamelCase ) return True def __snake_case ( self : int , lowerCamelCase : int ) -> int: if self.parents[disj_set] == disj_set: return disj_set __snake_case : Optional[int] = self.get_parent(self.parents[disj_set] ) return self.parents[disj_set]
81
1
# Algorithm for the pigeonhole sorting def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : Optional[Any] = min(__lowerCamelCase ) # min() finds the minimum value __snake_case : Any = max(__lowerCamelCase ) # max() finds the maximum value __snake_case : List[Any] = max_val - min_val + 1 # size is difference of max and min values plus one # list of pigeonholes of size equal to the variable size __snake_case : Optional[Any] = [0] * size # Populate the pigeonholes. for x in a: assert isinstance(__lowerCamelCase , __lowerCamelCase ), "integers only please" holes[x - min_val] += 1 # Putting the elements back into the array in an order. __snake_case : List[Any] = 0 for count in range(__lowerCamelCase ): while holes[count] > 0: holes[count] -= 1 __snake_case : Optional[Any] = count + min_val i += 1 def lowerCAmelCase_ ( ): __snake_case : int = [8, 3, 2, 7, 4, 6, 8] pigeonhole_sort(__lowerCamelCase ) print("Sorted order is:" , " ".join(__lowerCamelCase ) ) if __name__ == "__main__": main()
81
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class a (unittest.TestCase ): """simple docstring""" def __init__( self : int , lowerCamelCase : Any , lowerCamelCase : int=7 , lowerCamelCase : str=3 , lowerCamelCase : Optional[int]=30 , lowerCamelCase : Dict=400 , lowerCamelCase : str=True , lowerCamelCase : str=None , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[Any]=[0.5, 0.5, 0.5] , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=1 / 255 , lowerCamelCase : Any=True , ) -> str: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p __snake_case : Optional[Any] = size if size is not None else {"shortest_edge": 18, "longest_edge": 1333} __snake_case : Optional[Any] = parent __snake_case : List[Any] = batch_size __snake_case : Optional[int] = num_channels __snake_case : str = min_resolution __snake_case : int = max_resolution __snake_case : int = do_resize __snake_case : Tuple = size __snake_case : Any = do_normalize __snake_case : int = image_mean __snake_case : Tuple = image_std __snake_case : Dict = do_rescale __snake_case : Optional[Any] = rescale_factor __snake_case : str = do_pad def __snake_case ( self : Any ) -> int: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def __snake_case ( self : Dict , lowerCamelCase : List[Any] , lowerCamelCase : Optional[Any]=False ) -> List[str]: if not batched: __snake_case : Dict = image_inputs[0] if isinstance(lowerCamelCase , Image.Image ): __snake_case , __snake_case : Dict = image.size else: __snake_case , __snake_case : List[str] = image.shape[1], image.shape[2] if w < h: __snake_case : Optional[int] = int(self.size["shortest_edge"] * h / w ) __snake_case : int = self.size["shortest_edge"] elif w > h: __snake_case : List[str] = self.size["shortest_edge"] __snake_case : Optional[Any] = int(self.size["shortest_edge"] * w / h ) else: __snake_case : List[Any] = self.size["shortest_edge"] __snake_case : Any = self.size["shortest_edge"] else: __snake_case : int = [] for image in image_inputs: __snake_case , __snake_case : List[Any] = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[0] )[0] __snake_case : str = max(lowerCamelCase , key=lambda lowerCamelCase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : str = ConditionalDetrImageProcessor if is_vision_available() else None def __snake_case ( self : Optional[int] ) -> Optional[int]: __snake_case : Optional[Any] = ConditionalDetrImageProcessingTester(self ) @property def __snake_case ( self : Any ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def __snake_case ( self : Optional[Any] ) -> Optional[int]: __snake_case : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase , "image_mean" ) ) self.assertTrue(hasattr(lowerCamelCase , "image_std" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_normalize" ) ) self.assertTrue(hasattr(lowerCamelCase , "do_resize" ) ) self.assertTrue(hasattr(lowerCamelCase , "size" ) ) def __snake_case ( self : Any ) -> Dict: __snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 18, "longest_edge": 1333} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) __snake_case : str = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowerCamelCase ) self.assertEqual(image_processor.size , {"shortest_edge": 42, "longest_edge": 84} ) self.assertEqual(image_processor.do_pad , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> Dict: pass def __snake_case ( self : Tuple ) -> str: # Initialize image_processing __snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , Image.Image ) # Test not batched input __snake_case : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) __snake_case : Dict = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> str: # Initialize image_processing __snake_case : Optional[int] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case : Any = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , numpify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , np.ndarray ) # Test not batched input __snake_case : Union[str, Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : List[Any] = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Optional[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def __snake_case ( self : int ) -> List[str]: # Initialize image_processing __snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase , torchify=lowerCamelCase ) for image in image_inputs: self.assertIsInstance(lowerCamelCase , torch.Tensor ) # Test not batched input __snake_case : Optional[Any] = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values __snake_case , __snake_case : List[Any] = self.image_processor_tester.get_expected_values(lowerCamelCase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched __snake_case : int = image_processing(lowerCamelCase , return_tensors="pt" ).pixel_values __snake_case , __snake_case : Union[str, Any] = self.image_processor_tester.get_expected_values(lowerCamelCase , batched=lowerCamelCase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def __snake_case ( self : Any ) -> Optional[int]: # prepare image and target __snake_case : List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : List[Any] = {"image_id": 39769, "annotations": target} # encode them __snake_case : List[str] = ConditionalDetrImageProcessor.from_pretrained("microsoft/conditional-detr-resnet-50" ) __snake_case : List[str] = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : Tuple = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : List[Any] = torch.tensor([58_87.96_00, 1_12_50.20_61, 48_93_53.84_38, 83_71_22.75_00, 14_79_67.51_56, 16_57_32.34_38] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : Optional[int] = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : List[Any] = torch.tensor([0.55_03, 0.27_65, 0.06_04, 0.22_15] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Optional[Any] = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : List[Any] = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : Dict = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify orig_size __snake_case : int = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Tuple = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) ) @slow def __snake_case ( self : str ) -> Tuple: # prepare image, target and masks_path __snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) with open("./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt" , "r" ) as f: __snake_case : str = json.loads(f.read() ) __snake_case : str = {"file_name": "000000039769.png", "image_id": 39769, "segments_info": target} __snake_case : Optional[int] = pathlib.Path("./tests/fixtures/tests_samples/COCO/coco_panoptic" ) # encode them __snake_case : int = ConditionalDetrImageProcessor(format="coco_panoptic" ) __snake_case : str = image_processing(images=lowerCamelCase , annotations=lowerCamelCase , masks_path=lowerCamelCase , return_tensors="pt" ) # verify pixel values __snake_case : List[str] = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding["pixel_values"].shape , lowerCamelCase ) __snake_case : Dict = torch.tensor([0.27_96, 0.31_38, 0.34_81] ) self.assertTrue(torch.allclose(encoding["pixel_values"][0, 0, 0, :3] , lowerCamelCase , atol=1E-4 ) ) # verify area __snake_case : Any = torch.tensor([14_79_79.68_75, 16_55_27.04_69, 48_46_38.59_38, 1_12_92.93_75, 58_79.65_62, 76_34.11_47] ) self.assertTrue(torch.allclose(encoding["labels"][0]["area"] , lowerCamelCase ) ) # verify boxes __snake_case : str = torch.Size([6, 4] ) self.assertEqual(encoding["labels"][0]["boxes"].shape , lowerCamelCase ) __snake_case : Optional[Any] = torch.tensor([0.26_25, 0.54_37, 0.46_88, 0.86_25] ) self.assertTrue(torch.allclose(encoding["labels"][0]["boxes"][0] , lowerCamelCase , atol=1E-3 ) ) # verify image_id __snake_case : Tuple = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding["labels"][0]["image_id"] , lowerCamelCase ) ) # verify is_crowd __snake_case : Dict = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding["labels"][0]["iscrowd"] , lowerCamelCase ) ) # verify class_labels __snake_case : int = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding["labels"][0]["class_labels"] , lowerCamelCase ) ) # verify masks __snake_case : List[Any] = 822873 self.assertEqual(encoding["labels"][0]["masks"].sum().item() , lowerCamelCase ) # verify orig_size __snake_case : List[str] = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding["labels"][0]["orig_size"] , lowerCamelCase ) ) # verify size __snake_case : Any = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding["labels"][0]["size"] , lowerCamelCase ) )
81
1
import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a : """simple docstring""" def __init__( self : Union[str, Any] , lowerCamelCase : Any , lowerCamelCase : Any=13 , lowerCamelCase : Union[str, Any]=32 , lowerCamelCase : Tuple=3 , lowerCamelCase : Any=4 , lowerCamelCase : Optional[int]=[10, 20, 30, 40] , lowerCamelCase : int=[2, 2, 3, 2] , lowerCamelCase : str=True , lowerCamelCase : Any=True , lowerCamelCase : Union[str, Any]=37 , lowerCamelCase : List[str]="gelu" , lowerCamelCase : Optional[int]=10 , lowerCamelCase : Optional[Any]=0.02 , lowerCamelCase : List[Any]=["stage2", "stage3", "stage4"] , lowerCamelCase : Union[str, Any]=[2, 3, 4] , lowerCamelCase : Dict=None , ) -> List[str]: __snake_case : int = parent __snake_case : int = batch_size __snake_case : List[Any] = image_size __snake_case : Any = num_channels __snake_case : Optional[int] = num_stages __snake_case : Union[str, Any] = hidden_sizes __snake_case : List[str] = depths __snake_case : Any = is_training __snake_case : Tuple = use_labels __snake_case : Dict = intermediate_size __snake_case : Any = hidden_act __snake_case : Optional[Any] = num_labels __snake_case : int = initializer_range __snake_case : Any = out_features __snake_case : int = out_indices __snake_case : Optional[Any] = scope def __snake_case ( self : Optional[int] ) -> List[str]: __snake_case : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Any = None if self.use_labels: __snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : str = self.get_config() return config, pixel_values, labels def __snake_case ( self : str ) -> List[Any]: return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def __snake_case ( self : Tuple , lowerCamelCase : Optional[Any] , lowerCamelCase : Tuple , lowerCamelCase : str ) -> Any: __snake_case : Any = ConvNextVaModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Optional[Any] = model(lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __snake_case ( self : List[str] , lowerCamelCase : List[str] , lowerCamelCase : Tuple , lowerCamelCase : str ) -> List[str]: __snake_case : Any = ConvNextVaForImageClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : int = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : List[Any] , lowerCamelCase : str , lowerCamelCase : Dict , lowerCamelCase : List[Any] ) -> Dict: __snake_case : Optional[int] = ConvNextVaBackbone(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Optional[Any] = model(lowerCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __snake_case : Union[str, Any] = None __snake_case : Optional[int] = ConvNextVaBackbone(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() __snake_case : Optional[Any] = model(lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __snake_case ( self : List[str] ) -> Union[str, Any]: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case : List[str] = config_and_inputs __snake_case : Optional[int] = {"pixel_values": pixel_values} return config, inputs_dict def __snake_case ( self : str ) -> Union[str, Any]: __snake_case : str = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case : int = config_and_inputs __snake_case : List[Any] = {"pixel_values": pixel_values, "labels": labels} return config, inputs_dict @require_torch class a (_lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Any = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) __UpperCAmelCase : Union[str, Any] = ( {"feature-extraction": ConvNextVaModel, "image-classification": ConvNextVaForImageClassification} if is_torch_available() else {} ) __UpperCAmelCase : str = False __UpperCAmelCase : List[str] = False __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : List[Any] = False __UpperCAmelCase : Optional[int] = False def __snake_case ( self : Optional[int] ) -> List[Any]: __snake_case : Union[str, Any] = ConvNextVaModelTester(self ) __snake_case : Any = ConfigTester(self , config_class=lowerCamelCase , has_text_modality=lowerCamelCase , hidden_size=37 ) def __snake_case ( self : Optional[Any] ) -> Optional[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __snake_case ( self : str ) -> Dict: return @unittest.skip(reason="ConvNextV2 does not use inputs_embeds" ) def __snake_case ( self : str ) -> str: pass @unittest.skip(reason="ConvNextV2 does not support input and output embeddings" ) def __snake_case ( self : Optional[int] ) -> str: pass @unittest.skip(reason="ConvNextV2 does not use feedforward chunking" ) def __snake_case ( self : Tuple ) -> Tuple: pass def __snake_case ( self : Tuple ) -> Optional[int]: if not self.model_tester.is_training: return for model_class in self.all_model_classes: __snake_case , __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_with_labels() __snake_case : Optional[int] = True if model_class.__name__ in [ *get_values(lowerCamelCase ), *get_values(lowerCamelCase ), ]: continue __snake_case : Optional[int] = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.train() __snake_case : List[Any] = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) __snake_case : List[Any] = model(**lowerCamelCase ).loss loss.backward() def __snake_case ( self : str ) -> Optional[Any]: if not self.model_tester.is_training: return for model_class in self.all_model_classes: __snake_case , __snake_case : str = self.model_tester.prepare_config_and_inputs_with_labels() __snake_case : Tuple = False __snake_case : List[str] = True if ( model_class.__name__ in [*get_values(lowerCamelCase ), *get_values(lowerCamelCase )] or not model_class.supports_gradient_checkpointing ): continue __snake_case : Dict = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.gradient_checkpointing_enable() model.train() __snake_case : int = self._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) __snake_case : List[str] = model(**lowerCamelCase ).loss loss.backward() def __snake_case ( self : Optional[int] ) -> Tuple: __snake_case , __snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Dict = model_class(lowerCamelCase ) __snake_case : List[str] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Dict = [*signature.parameters.keys()] __snake_case : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase ) def __snake_case ( self : Optional[Any] ) -> str: __snake_case : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase ) def __snake_case ( self : Dict ) -> Dict: def check_hidden_states_output(lowerCamelCase : Any , lowerCamelCase : List[str] , lowerCamelCase : List[Any] ): __snake_case : str = model_class(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() with torch.no_grad(): __snake_case : Tuple = model(**self._prepare_for_class(lowerCamelCase , lowerCamelCase ) ) __snake_case : Union[str, Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __snake_case : int = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __snake_case , __snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Optional[Any] = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : Dict = True check_hidden_states_output(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def __snake_case ( self : int ) -> str: __snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase ) @slow def __snake_case ( self : Tuple ) -> str: for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : int = ConvNextVaModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) def lowerCAmelCase_ ( ): __snake_case : Union[str, Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a (unittest.TestCase ): """simple docstring""" @cached_property def __snake_case ( self : Optional[int] ) -> List[str]: return AutoImageProcessor.from_pretrained("facebook/convnextv2-tiny-1k-224" ) if is_vision_available() else None @slow def __snake_case ( self : int ) -> List[str]: __snake_case : int = ConvNextVaForImageClassification.from_pretrained("facebook/convnextv2-tiny-1k-224" ).to(lowerCamelCase ) __snake_case : Union[str, Any] = self.default_image_processor __snake_case : Optional[Any] = prepare_img() __snake_case : List[Any] = preprocessor(images=lowerCamelCase , return_tensors="pt" ).to(lowerCamelCase ) # forward pass with torch.no_grad(): __snake_case : int = model(**lowerCamelCase ) # verify the logits __snake_case : List[str] = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowerCamelCase ) __snake_case : Optional[int] = torch.tensor([0.99_96, 0.19_66, -0.43_86] ).to(lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase , atol=1E-4 ) )
81
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig _snake_case : Optional[Any] = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring _snake_case : Dict = "UperNetConfig" class a (nn.Module ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : Union[int, Tuple[int, int]] , lowerCamelCase : Union[int, Tuple[int, int], str] = 0 , lowerCamelCase : bool = False , lowerCamelCase : Union[int, Tuple[int, int]] = 1 , ) -> None: super().__init__() __snake_case : Union[str, Any] = nn.Convad( in_channels=lowerCamelCase , out_channels=lowerCamelCase , kernel_size=lowerCamelCase , padding=lowerCamelCase , bias=lowerCamelCase , dilation=lowerCamelCase , ) __snake_case : Dict = nn.BatchNormad(lowerCamelCase ) __snake_case : List[Any] = nn.ReLU() def __snake_case ( self : List[Any] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : Dict = self.conv(lowerCamelCase ) __snake_case : int = self.batch_norm(lowerCamelCase ) __snake_case : Optional[Any] = self.activation(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : str , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int ) -> None: super().__init__() __snake_case : Tuple = [ nn.AdaptiveAvgPoolad(lowerCamelCase ), UperNetConvModule(lowerCamelCase , lowerCamelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : Dict , lowerCamelCase : torch.Tensor ) -> torch.Tensor: __snake_case : List[str] = input for layer in self.layers: __snake_case : Tuple = layer(lowerCamelCase ) return hidden_state class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Tuple[int, ...] , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : bool ) -> None: super().__init__() __snake_case : Dict = pool_scales __snake_case : List[str] = align_corners __snake_case : List[Any] = in_channels __snake_case : str = channels __snake_case : Optional[Any] = [] for i, pool_scale in enumerate(lowerCamelCase ): __snake_case : Union[str, Any] = UperNetPyramidPoolingBlock(pool_scale=lowerCamelCase , in_channels=lowerCamelCase , channels=lowerCamelCase ) self.blocks.append(lowerCamelCase ) self.add_module(str(lowerCamelCase ) , lowerCamelCase ) def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> List[torch.Tensor]: __snake_case : Tuple = [] for ppm in self.blocks: __snake_case : Any = ppm(lowerCamelCase ) __snake_case : List[Any] = nn.functional.interpolate( lowerCamelCase , size=x.size()[2:] , mode="bilinear" , align_corners=self.align_corners ) ppm_outs.append(lowerCamelCase ) return ppm_outs class a (nn.Module ): """simple docstring""" def __init__( self : int , lowerCamelCase : str , lowerCamelCase : str ) -> List[str]: super().__init__() __snake_case : Dict = config __snake_case : List[str] = config.pool_scales # e.g. (1, 2, 3, 6) __snake_case : Tuple = in_channels __snake_case : str = config.hidden_size __snake_case : List[str] = False __snake_case : Any = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module __snake_case : Tuple = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) __snake_case : List[str] = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module __snake_case : List[Any] = nn.ModuleList() __snake_case : Dict = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer __snake_case : Union[str, Any] = UperNetConvModule(lowerCamelCase , self.channels , kernel_size=1 ) __snake_case : Optional[int] = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(lowerCamelCase ) self.fpn_convs.append(lowerCamelCase ) __snake_case : int = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def __snake_case ( self : List[str] ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : List[Any] ) -> str: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : List[Any] , lowerCamelCase : Tuple ) -> Optional[int]: __snake_case : str = inputs[-1] __snake_case : int = [x] psp_outs.extend(self.psp_modules(lowerCamelCase ) ) __snake_case : Tuple = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Union[str, Any] = self.bottleneck(lowerCamelCase ) return output def __snake_case ( self : int , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # build laterals __snake_case : Any = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(lowerCamelCase ) ) # build top-down path __snake_case : Dict = len(lowerCamelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Union[str, Any] = laterals[i - 1].shape[2:] __snake_case : Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=lowerCamelCase , mode="bilinear" , align_corners=self.align_corners ) # build outputs __snake_case : str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): __snake_case : Tuple = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode="bilinear" , align_corners=self.align_corners ) __snake_case : str = torch.cat(lowerCamelCase , dim=1 ) __snake_case : Optional[Any] = self.fpn_bottleneck(lowerCamelCase ) __snake_case : Tuple = self.classifier(lowerCamelCase ) return output class a (nn.Module ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Optional[int] , lowerCamelCase : int = 2 , lowerCamelCase : int = 3 , lowerCamelCase : Union[int, Tuple[int, int]] = 1 ) -> None: super().__init__() __snake_case : List[Any] = config __snake_case : List[str] = config.auxiliary_in_channels __snake_case : List[Any] = config.auxiliary_channels __snake_case : Tuple = config.auxiliary_num_convs __snake_case : int = config.auxiliary_concat_input __snake_case : Optional[int] = in_index __snake_case : Tuple = (kernel_size // 2) * dilation __snake_case : Optional[int] = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=lowerCamelCase , padding=lowerCamelCase , dilation=lowerCamelCase ) ) if self.num_convs == 0: __snake_case : Union[str, Any] = nn.Identity() else: __snake_case : Any = nn.Sequential(*lowerCamelCase ) if self.concat_input: __snake_case : int = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=lowerCamelCase , padding=kernel_size // 2 ) __snake_case : Optional[Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def __snake_case ( self : Dict ) -> Optional[Any]: self.apply(self._init_weights ) def __snake_case ( self : Tuple , lowerCamelCase : Tuple ) -> Optional[int]: if isinstance(lowerCamelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def __snake_case ( self : Optional[int] , lowerCamelCase : torch.Tensor ) -> torch.Tensor: # just take the relevant feature maps __snake_case : List[str] = encoder_hidden_states[self.in_index] __snake_case : Optional[Any] = self.convs(lowerCamelCase ) if self.concat_input: __snake_case : Optional[Any] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) __snake_case : Union[str, Any] = self.classifier(lowerCamelCase ) return output class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Optional[Any] = UperNetConfig __UpperCAmelCase : int = "pixel_values" __UpperCAmelCase : str = True def __snake_case ( self : Union[str, Any] , lowerCamelCase : List[Any] ) -> List[Any]: if isinstance(lowerCamelCase , lowerCamelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def __snake_case ( self : Optional[Any] ) -> List[str]: self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def __snake_case ( self : List[Any] , lowerCamelCase : Any , lowerCamelCase : Optional[Any]=False ) -> Dict: if isinstance(lowerCamelCase , lowerCamelCase ): __snake_case : Union[str, Any] = value _snake_case : Dict = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" _snake_case : Tuple = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." , _lowerCAmelCase , ) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Dict , lowerCamelCase : int ) -> Optional[int]: super().__init__(lowerCamelCase ) __snake_case : Any = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) __snake_case : Union[str, Any] = UperNetHead(lowerCamelCase , in_channels=self.backbone.channels ) __snake_case : Any = UperNetFCNHead(lowerCamelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length" ) ) @replace_return_docstrings(output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC ) def __snake_case ( self : Union[str, Any] , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[torch.Tensor] = None , lowerCamelCase : Optional[bool] = None , ) -> Union[tuple, SemanticSegmenterOutput]: __snake_case : Any = return_dict if return_dict is not None else self.config.use_return_dict __snake_case : int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __snake_case : str = output_attentions if output_attentions is not None else self.config.output_attentions __snake_case : Tuple = self.backbone.forward_with_filtered_kwargs( lowerCamelCase , output_hidden_states=lowerCamelCase , output_attentions=lowerCamelCase ) __snake_case : List[Any] = outputs.feature_maps __snake_case : List[Any] = self.decode_head(lowerCamelCase ) __snake_case : List[str] = nn.functional.interpolate(lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : Optional[int] = None if self.auxiliary_head is not None: __snake_case : Dict = self.auxiliary_head(lowerCamelCase ) __snake_case : Dict = nn.functional.interpolate( lowerCamelCase , size=pixel_values.shape[2:] , mode="bilinear" , align_corners=lowerCamelCase ) __snake_case : int = None if labels is not None: if self.config.num_labels == 1: raise ValueError("The number of labels should be greater than one" ) else: # compute weighted loss __snake_case : Any = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) __snake_case : Union[str, Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = loss_fct(lowerCamelCase , lowerCamelCase ) __snake_case : Optional[Any] = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: __snake_case : Any = (logits,) + outputs[1:] else: __snake_case : str = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
81
1
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor _snake_case : Optional[int] = logging.get_logger(__name__) class a (_lowerCAmelCase ): """simple docstring""" def __init__( self : Optional[Any] , *lowerCamelCase : Any , **lowerCamelCase : Union[str, Any] ) -> None: warnings.warn( "The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use FlavaImageProcessor instead." , lowerCamelCase , ) super().__init__(*lowerCamelCase , **lowerCamelCase )
81
from __future__ import annotations def lowerCAmelCase_ ( __lowerCamelCase ): if not nums: return 0 __snake_case : Any = nums[0] __snake_case : str = 0 for num in nums[1:]: __snake_case , __snake_case : List[str] = ( max_excluding + num, max(__lowerCamelCase , __lowerCamelCase ), ) return max(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
81
1
from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : Tuple = logging.get_logger(__name__) _snake_case : Tuple = { "huggingface/informer-tourism-monthly": ( "https://huggingface.co/huggingface/informer-tourism-monthly/resolve/main/config.json" ), # See all Informer models at https://huggingface.co/models?filter=informer } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : str = "informer" __UpperCAmelCase : Optional[Any] = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", "num_hidden_layers": "encoder_layers", } def __init__( self : int , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[int] = None , lowerCamelCase : str = "student_t" , lowerCamelCase : str = "nll" , lowerCamelCase : int = 1 , lowerCamelCase : List[int] = None , lowerCamelCase : Optional[Union[str, bool]] = "mean" , lowerCamelCase : int = 0 , lowerCamelCase : int = 0 , lowerCamelCase : int = 0 , lowerCamelCase : int = 0 , lowerCamelCase : Optional[List[int]] = None , lowerCamelCase : Optional[List[int]] = None , lowerCamelCase : int = 64 , lowerCamelCase : int = 32 , lowerCamelCase : int = 32 , lowerCamelCase : int = 2 , lowerCamelCase : int = 2 , lowerCamelCase : int = 2 , lowerCamelCase : int = 2 , lowerCamelCase : bool = True , lowerCamelCase : str = "gelu" , lowerCamelCase : float = 0.05 , lowerCamelCase : float = 0.1 , lowerCamelCase : float = 0.1 , lowerCamelCase : float = 0.1 , lowerCamelCase : float = 0.1 , lowerCamelCase : int = 100 , lowerCamelCase : float = 0.02 , lowerCamelCase : Optional[Any]=True , lowerCamelCase : str = "prob" , lowerCamelCase : int = 5 , lowerCamelCase : bool = True , **lowerCamelCase : str , ) -> str: # time series specific configuration __snake_case : Any = prediction_length __snake_case : str = context_length or prediction_length __snake_case : Any = distribution_output __snake_case : Optional[Any] = loss __snake_case : List[Any] = input_size __snake_case : List[Any] = num_time_features __snake_case : Union[str, Any] = lags_sequence if lags_sequence is not None else [1, 2, 3, 4, 5, 6, 7] __snake_case : Tuple = scaling __snake_case : List[Any] = num_dynamic_real_features __snake_case : Tuple = num_static_real_features __snake_case : Union[str, Any] = num_static_categorical_features # set cardinality if cardinality and num_static_categorical_features > 0: if len(lowerCamelCase ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case : Tuple = cardinality else: __snake_case : Tuple = [0] # set embedding_dimension if embedding_dimension and num_static_categorical_features > 0: if len(lowerCamelCase ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case : Tuple = embedding_dimension else: __snake_case : Union[str, Any] = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case : List[str] = num_parallel_samples # Transformer architecture configuration __snake_case : Optional[int] = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case : str = d_model __snake_case : int = encoder_attention_heads __snake_case : Any = decoder_attention_heads __snake_case : Any = encoder_ffn_dim __snake_case : List[Any] = decoder_ffn_dim __snake_case : Dict = encoder_layers __snake_case : Dict = decoder_layers __snake_case : Union[str, Any] = dropout __snake_case : Tuple = attention_dropout __snake_case : int = activation_dropout __snake_case : List[str] = encoder_layerdrop __snake_case : Optional[int] = decoder_layerdrop __snake_case : List[str] = activation_function __snake_case : Dict = init_std __snake_case : Union[str, Any] = use_cache # Informer __snake_case : Optional[Any] = attention_type __snake_case : List[Any] = sampling_factor __snake_case : Optional[int] = distil super().__init__(is_encoder_decoder=lowerCamelCase , **lowerCamelCase ) @property def __snake_case ( self : Optional[Any] ) -> int: return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
81
from __future__ import annotations from typing import Any def lowerCAmelCase_ ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase , [] , 0 ) def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase , __lowerCamelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": _snake_case : list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
81
1